/*------------------------------------
Iwa_PerspectiveDistortFx
奥行き方向に台形歪みを行うエフェクト
ディティールを保持するため、引き伸ばす量に応じて素材の解像度を上げる
------------------------------------*/
#include "iwa_perspectivedistortfx.h"
#include "tparamuiconcept.h"
// #include "ino_common.h"
#include "trop.h"
namespace {
inline float4 getSource_CPU(float4 *source_host, TDimensionI &dim, int pos_x,
int pos_y) {
if (pos_x < 0 || pos_x >= dim.lx || pos_y < 0 || pos_y >= dim.ly)
return float4{0.0f, 0.0f, 0.0f, 0.0f};
return source_host[pos_y * dim.lx + pos_x];
}
inline float4 interp_CPU(float4 val1, float4 val2, float ratio) {
float4 ret;
ret.x = (1.0f - ratio) * val1.x + ratio * val2.x;
ret.y = (1.0f - ratio) * val1.y + ratio * val2.y;
ret.z = (1.0f - ratio) * val1.z + ratio * val2.z;
ret.w = (1.0f - ratio) * val1.w + ratio * val2.w;
return ret;
}
} // namespace
/*------------------------------------------------------------
出力結果をChannel値に変換して格納
------------------------------------------------------------*/
template <typename RASTER, typename PIXEL>
void Iwa_PerspectiveDistortFx::setOutputRaster(float4 *srcMem,
const RASTER dstRas,
TDimensionI dim, int drawLevel) {
dstRas->fill(PIXEL::Transparent);
float4 *chan_p = srcMem;
for (int j = 0; j < drawLevel; j++) {
if (j >= dstRas->getLy()) break;
PIXEL *pix = dstRas->pixels(j);
for (int i = 0; i < dstRas->getLx(); i++, chan_p++, pix++) {
float val;
val = (*chan_p).x * (float)PIXEL::maxChannelValue + 0.5f;
pix->r = (typename PIXEL::Channel)((val > (float)PIXEL::maxChannelValue)
? (float)PIXEL::maxChannelValue
: val);
val = (*chan_p).y * (float)PIXEL::maxChannelValue + 0.5f;
pix->g = (typename PIXEL::Channel)((val > (float)PIXEL::maxChannelValue)
? (float)PIXEL::maxChannelValue
: val);
val = (*chan_p).z * (float)PIXEL::maxChannelValue + 0.5f;
pix->b = (typename PIXEL::Channel)((val > (float)PIXEL::maxChannelValue)
? (float)PIXEL::maxChannelValue
: val);
val = (*chan_p).w * (float)PIXEL::maxChannelValue + 0.5f;
pix->m = (typename PIXEL::Channel)((val > (float)PIXEL::maxChannelValue)
? (float)PIXEL::maxChannelValue
: val);
}
}
}
template <>
void Iwa_PerspectiveDistortFx::setOutputRaster<TRasterFP, TPixelF>(
float4 *srcMem, const TRasterFP dstRas, TDimensionI dim, int drawLevel) {
dstRas->fill(TPixelF::Transparent);
float4 *chan_p = srcMem;
for (int j = 0; j < drawLevel; j++) {
if (j >= dstRas->getLy()) break;
TPixelF *pix = dstRas->pixels(j);
for (int i = 0; i < dstRas->getLx(); i++, chan_p++, pix++) {
pix->r = (*chan_p).x;
pix->g = (*chan_p).y;
pix->b = (*chan_p).z;
pix->m = (*chan_p).w;
}
}
}
/*------------------------------------------------------------
ソース画像を0〜1に正規化してホストメモリに読み込む
------------------------------------------------------------*/
template <typename RASTER, typename PIXEL>
void Iwa_PerspectiveDistortFx::setSourceRaster(const RASTER srcRas,
float4 *dstMem,
TDimensionI dim) {
float4 *chann_p = dstMem;
for (int j = 0; j < dim.ly; j++) {
PIXEL *pix = srcRas->pixels(j);
for (int i = 0; i < dim.lx; i++, pix++, chann_p++) {
(*chann_p).x = (float)pix->r / (float)PIXEL::maxChannelValue;
(*chann_p).y = (float)pix->g / (float)PIXEL::maxChannelValue;
(*chann_p).z = (float)pix->b / (float)PIXEL::maxChannelValue;
(*chann_p).w = (float)pix->m / (float)PIXEL::maxChannelValue;
}
}
}
//------------------------------------------------------------
Iwa_PerspectiveDistortFx::Iwa_PerspectiveDistortFx()
: m_vanishingPoint(TPointD(0, 0))
, m_anchorPoint(TPointD(0, -100.0))
, m_precision(300.0 / 162.5) {
/*- 共通パラメータのバインド -*/
addInputPort("Source", m_source);
bindParam(this, "vanishingPoint", m_vanishingPoint);
bindParam(this, "anchorPoint", m_anchorPoint);
bindParam(this, "precision", m_precision);
m_vanishingPoint->getX()->setMeasureName("fxLength");
m_vanishingPoint->getY()->setMeasureName("fxLength");
m_anchorPoint->getX()->setMeasureName("fxLength");
m_anchorPoint->getY()->setMeasureName("fxLength");
m_precision->setValueRange(1.0, 2.0);
enableComputeInFloat(true);
}
//------------------------------------
bool Iwa_PerspectiveDistortFx::doGetBBox(double frame, TRectD &bBox,
const TRenderSettings &info) {
if (m_source.isConnected()) {
bool ret = m_source->doGetBBox(frame, bBox, info);
if (ret) bBox = TConsts::infiniteRectD;
return ret;
}
return false;
}
//------------------------------------
bool Iwa_PerspectiveDistortFx::canHandle(const TRenderSettings &info,
double frame) {
return false;
}
//------------------------------------
void Iwa_PerspectiveDistortFx::doCompute(TTile &tile, double frame,
const TRenderSettings &rend_sets) {
/*- ソース画像が刺さっていなければreturn -*/
if (!m_source.isConnected()) {
tile.getRaster()->clear();
return;
}
/*- パラメータの取得 -*/
TPointD vp = m_vanishingPoint->getValue(frame);
TPointD ap = m_anchorPoint->getValue(frame);
double precision = m_precision->getValue(frame);
/*- 基準点が消失点よりも上にあったらreturn -*/
if (vp.y <= ap.y) {
std::cout << "the anchor must be lower than the vanishing point"
<< std::endl;
tile.getRaster()->clear();
return;
}
TAffine aff = rend_sets.m_affine;
/*- カメラ左下を中心としたピクセル座標に変換する -*/
TPointD vanishingPoint = aff * vp -
(tile.m_pos + tile.getRaster()->getCenterD()) +
TPointD(rend_sets.m_cameraBox.getLx() / 2.0,
rend_sets.m_cameraBox.getLy() / 2.0);
TPointD anchorPoint = aff * ap -
(tile.m_pos + tile.getRaster()->getCenterD()) +
TPointD(rend_sets.m_cameraBox.getLx() / 2.0,
rend_sets.m_cameraBox.getLy() / 2.0);
double offs = vanishingPoint.x - rend_sets.m_cameraBox.getLx() / 2.0;
/*- 取り込む解像度の倍率 -*/
TRectD rectOut(tile.m_pos, TDimensionD(tile.getRaster()->getLx(),
tile.getRaster()->getLy()));
TDimensionI outDim(rectOut.getLx(), rectOut.getLy());
/*- ソース画像を正規化して格納 -*/
/*- 素材を拡大させて持ち込む -*/
TDimensionI sourceDim(rectOut.getLx() * (int)tceil(precision), anchorPoint.y);
float4 *source_host = nullptr;
TRenderSettings new_sets(rend_sets);
new_sets.m_affine *= TTranslation(vp.x, vp.y);
new_sets.m_affine *= TScale(precision, 1.0);
new_sets.m_affine *= TTranslation(-vp.x, -vp.y);
double sourcePosX = offs + precision * (tile.m_pos.x - offs);
TTile sourceTile;
m_source->allocateAndCompute(sourceTile, TPointD(sourcePosX, tile.m_pos.y),
sourceDim, tile.getRaster(), frame, new_sets);
TRasterGR8P source_host_ras;
if (tile.getRaster()->getPixelSize() == 4 ||
tile.getRaster()->getPixelSize() == 8) {
source_host_ras = TRasterGR8P(sourceDim.lx * sizeof(float4), sourceDim.ly);
source_host_ras->lock();
source_host = (float4 *)source_host_ras->getRawData();
/*- タイルの画像を0〜1に正規化してホストメモリに読み込む -*/
TRaster32P ras32 = (TRaster32P)sourceTile.getRaster();
TRaster64P ras64 = (TRaster64P)sourceTile.getRaster();
if (ras32)
setSourceRaster<TRaster32P, TPixel32>(ras32, source_host, sourceDim);
else if (ras64)
setSourceRaster<TRaster64P, TPixel64>(ras64, source_host, sourceDim);
} else if (tile.getRaster()->getPixelSize() == 16) {
source_host =
reinterpret_cast<float4 *>(sourceTile.getRaster()->getRawData());
} else {
throw TRopException("unsupported input pixel type");
}
TDimensionI resultDim(rectOut.getLx(), anchorPoint.y);
TRasterGR8P result_host_ras(resultDim.lx * sizeof(float4), resultDim.ly);
/*- 結果を収めるメモリ -*/
float4 *result_host;
result_host_ras->lock();
result_host = (float4 *)result_host_ras->getRawData();
doCompute_CPU(tile, frame, rend_sets, vanishingPoint, anchorPoint,
source_host, result_host, sourceDim, resultDim, precision,
offs);
if (source_host_ras) source_host_ras->unlock();
/*- 出力結果をChannel値に変換して格納 -*/
TRaster32P outRas32 = (TRaster32P)tile.getRaster();
TRaster64P outRas64 = (TRaster64P)tile.getRaster();
TRasterFP outRasF = (TRasterFP)tile.getRaster();
if (outRas32)
setOutputRaster<TRaster32P, TPixel32>(result_host, outRas32, outDim,
resultDim.ly);
else if (outRas64)
setOutputRaster<TRaster64P, TPixel64>(result_host, outRas64, outDim,
resultDim.ly);
else if (outRasF)
setOutputRaster<TRasterFP, TPixelF>(result_host, outRasF, outDim,
resultDim.ly);
result_host_ras->unlock();
}
//------------------------------------
void Iwa_PerspectiveDistortFx::doCompute_CPU(
TTile &tile, const double frame, const TRenderSettings &settings,
TPointD &vp, TPointD &ap, float4 *source_host, float4 *result_host,
TDimensionI &sourceDim, TDimensionI &resultDim, const double precision,
const double offs) {
/*- 結果ポインタ -*/
float4 *result_p = result_host;
/*- 結果画像内でループ -*/
for (int j = 0; j < resultDim.ly; j++) {
/*- サンプルスタート地点 -*/
double sampleX =
precision * (vp.x * (ap.y - (double)j) / (vp.y - (double)j));
/*- 1ピクセル横に移動した場合のサンプリング位置の移動量 -*/
double dx = precision * (vp.y - ap.y) / (vp.y - (double)j);
for (int i = 0; i < resultDim.lx; i++, result_p++, sampleX += dx) {
int index = (int)tfloor(sampleX);
double ratio = sampleX - (double)index;
(*result_p) = interp_CPU(
getSource_CPU(source_host, sourceDim, index, j),
getSource_CPU(source_host, sourceDim, index + 1, j), ratio);
}
}
}
//------------------------------------
void Iwa_PerspectiveDistortFx::getParamUIs(TParamUIConcept *&concepts,
int &length) {
concepts = new TParamUIConcept[length = 2];
concepts[0].m_type = TParamUIConcept::POINT;
concepts[0].m_label = "Vanishing Point";
concepts[0].m_params.push_back(m_vanishingPoint);
concepts[1].m_type = TParamUIConcept::POINT;
concepts[1].m_label = "Anchor Point";
concepts[1].m_params.push_back(m_anchorPoint);
}
//------------------------------------
bool Iwa_PerspectiveDistortFx::toBeComputedInLinearColorSpace(
bool settingsIsLinear, bool tileIsLinear) const {
return tileIsLinear;
}
FX_PLUGIN_IDENTIFIER(Iwa_PerspectiveDistortFx, "iwa_PerspectiveDistortFx")