AnalyzeHDRPLit
Lit
在改动Lit之前有必要详细的逐行分析一下Lit的Shader
虽然类似的文章已经有了,但还是选择自己手敲一遍以加深印象。
版本是14.0.8
Common.hlsl
路径 core/ShaderLibrary/Common.hlsl
untiy为SRP单独抽象出了一个库,com.unity.render-pipelines.core,包含了很多基础的渲染代码,同时被URP和HDRP使用。
约定
首先就规定了一些惯例
坐标系:
- 世界坐标系:Yup左手
- 当从世界空间到视图空间时,视图空间中的单位为右手,矩阵的行列式为负
- 对于立方体贴图采样(反射探针),视图空间仍然是左手(立方体贴图约定)并且行列式为正。
- 单位:一米
然后是变量的后缀的约定
- WS:世界
- RWS:相机相关的世界
- VS:视角
- OS:物体
- CS:齐次裁切
- TS:切线
- TXS:贴图
默认向量都是归一化的,非归一化用un前缀
用大写字母表示常见向量,向量总是指向像素外部。
大写字母也意味着归一化,除非有un的前缀
- V:视角
- L:光线
- N:法线
- H:半程
输入输出的结构体用帕斯卡命名,前缀表示类型
如 AttributesDefault VaryingsDefault
当使用这些结构体时,用input/output作为变量名
常量浮点写作1.0 不是1,1.0f,1.0h
全局变量uniform 使用_前缀,大写,然后驼峰 _LowercaseThenCamelCase
不要使用in,仅使用out或inout关键字,也不要使用inline
当声明out参数时,放到最后
其他shader库不应包含common.hlsl,应放在.shader文件中。
所以全局变量应放在常量缓冲区中。
在C#与hlsl中共享结构定义。
结构要基于float4对齐。
定义数组时,总是使用float4
不要使用 SetGlobalFloatArray 或 SetComputeFloatParams
这个库中的函数都是无状态的,即没有全局声明。
需要显式声明精度的可以使用float与half,同时支持两者的,可以使用real。
real
这里就是依据定义的宏来确定是real的精度,大致看下来在移动端以及switch上,默认会启用half精度。当然可以手动指定REAL_IS_HALF为1。
宏定义
这里定义了一些方便使用的宏
#define SV_POSITION_QUALIFIERS
#define DEPTH_OFFSET_SEMANTIC SV_Depth
然后依据宏选择对应的api
PC就对应了D3D11.hlsl,大致包含了一些采样贴图的方式,等会再回过头来看看这个文件吧
接下来包含了Macros.hlsl与Random.hlsl,顾名思义,前者是主要是一些常量和一些模板方法,后者就是随机数了。
然后是Vulkan支持的Native Renderpass。这里先跳过。
然后是为了支持光追定义的宏
#if (SHADER_STAGE_RAY_TRACING && UNITY_RAY_TRACING_GLOBAL_RESOURCES)
#define GLOBAL_RESOURCE(type, name, reg) type name : register(reg, space1);
#define GLOBAL_CBUFFER_START(name, reg) cbuffer name : register(reg, space1) {
#else
#define GLOBAL_RESOURCE(type, name, reg) type name;
#define GLOBAL_CBUFFER_START(name, reg) CBUFFER_START(name)
#endif
工具函数
再往下就是存储的工具函数了。、
- 数学函数,如remap01,min,max,HasFlag等。
- 采样贴图
- 深度相关
- 还包含了一个PositionInputs的结构体,
- 并实现了一系列空间变换相关的函数。
D3D.hlsl
路径 core/ShaderLibrary/API/D3D.hlsl
是放到API中的,同级别的还有PSSL、Switch、Vulkan等等。所以可以猜出应该类似与接口一样,实现Unity常见的特定于平台的函数。
常量
首先是用宏定义了一些常量。
如下
#define UNITY_UV_STARTS_AT_TOP 1
#define UNITY_REVERSED_Z 1
#define UNITY_NEAR_CLIP_VALUE (1.0)
// This value will not go through any matrix projection conversion
#define UNITY_RAW_FAR_CLIP_VALUE (0.0)
#define VERTEXID_SEMANTIC SV_VertexID
#define INSTANCEID_SEMANTIC SV_InstanceID
#define FRONT_FACE_SEMANTIC SV_IsFrontFace
#define FRONT_FACE_TYPE bool
#define IS_FRONT_VFACE(VAL, FRONT, BACK) ((VAL) ? (FRONT) : (BACK))
都可以通过名称来看出含义。
支持的特性
#define PLATFORM_SUPPORTS_EXPLICIT_BINDING
#define PLATFORM_NEEDS_UNORM_UAV_SPECIFIER
#define PLATFORM_SUPPORTS_BUFFER_ATOMICS_IN_PIXEL_SHADER
#define PLATFORM_SUPPORTS_PRIMITIVE_ID_IN_PIXEL_SHADER
采样贴图
这里就顺便看下各种采样的区别吧
Sample 采样
SampleLevel 指定mipmap级别
SampleBias 输入mipmap偏差
SampleGrad 用梯度来选择mip级别
SampleCmpLevelZero 采样后与0比较
以及
Load 仅加载,即不寻址不过滤,使用指定的lod
Gather 收集周围4个像素的R通道
ShaderVariables.hlsl
主要是定义一个名为UnityPerDraw的CBuffer。包含了和材质相关的一些全局变量,如球谐函数参数,renderBox,光照贴图参数,渲染层等等和材质相关的。
还顺便声明了一系列常用的采样器。
再往下就是贴图的声明,如深度,光照,阴影,ProbeVolume。
这里还改了SAMPLE_TEXTURE2D与SAMPLE_TEXTURE2D_BIAS的宏,全局应用_GlobalMipBias
然后是采样这些贴图的工具函数 如
SampleCameraDepth、SampleCameraColor等
接着还是一些关于相机矩阵位置等参数的工具函数,如
GetRawUnityObjectToWorld、GetCameraPositionWS等等。
TextureXR.hlsl
顾名思义,是为了XR的单通道渲染定义的一系列在XR中采样贴图的宏
ShaderVariablesGlobal.cs.hlsl
这是一个由cs代码来自动生成的库文件。
单纯用来定义了一些全局变量,如_ViewMatrix、_ScreenSize、_Time等和材质无关的属性。
由此可以取看看ShaderVariablesGlobal.cs,通过cs代码来生成hlsl在HDRP中很常见。
ShaderVariablesGlobal.cs
这个文件主要就是定义了一个结构体,其内容和hlsl中的cbuffer一模一样,值得注意的是这里的参数的排布应用了float4对齐。
FragInputs.hlsl
一个结构体FragInputs,片元的输入,但这并不是Vert或者Frag的输出和输入,因为unity在传输时会打包再解包。
LitProperties.cs
和材质相关的属性,各种贴图与参数
LightLoopDef.hlsl
光照计算相关的定义
Lit.hlsl
PBR BSDF相关的计算
LightLoop.hlsl
?计算全部光源并汇总?
VaryingMesh.hlsl
顶点与片元的输入与输出,也包含了打包与解包相关函数。
LitData.hlsl
???
ShaderPassForward.hlsl
这里定义了前向渲染的顶点和片元着色器函数Vert与Frag。
Vert
这里倒是意外的简单,常规的坐标变换,打包输出到片元着色器。
Frag
头大的来了。
一行一行看!!!
函数输入输出
输入就是从顶点打包过来的。
这里根据几个宏来判断是否需要多Target输出,
共有虚拟纹理,分光?,运动向量,深度偏移。这些都先不考虑,只看outColor。
FragInputs input = UnpackVaryingsToFragInputs(packedInput);
解包数据
AdjustFragInputsToOffScreenRendering(input, _OffScreenRendering > 0,
_OffScreenDownsampleFactor);
离屏渲染??
目前没有搞明白有什么作用,大致看起来是用来处理低分辨率的。
uint2 tileIndex = uint2(input.positionSS.xy) / GetTileSize();
获取像素网格坐标int
PositionInputs posInput = GetPositionInput(input.positionSS.xy,
_ScreenSize.zw, input.positionSS.z, input.positionSS.w,
input.positionRWS.xyz, tileIndex);
struct PositionInputs
{
float3 positionWS; // World space position (could be camera-relative)
float2 positionNDC; // Normalized screen coordinates within the viewport : [0, 1) (with the half-pixel offset)
uint2 positionSS; // Screen space pixel coordinates : [0, NumPixels)
uint2 tileCoord; // Screen tile coordinates : [0, NumTiles)
float deviceDepth; // Depth from the depth buffer : [0, 1] (typically reversed)
float linearDepth; // View space Z coordinate : [Near, Far]
};
获取一系列各种空间下的坐标
float3 V = GetWorldSpaceNormalizeViewDir(input.positionRWS);
计算视线
SurfaceData surfaceData;
BuiltinData builtinData;
GetSurfaceAndBuiltinData(input, V, posInput, surfaceData,
builtinData);
这里就为之后的光照计算准备数据了,来源就是采样各个贴图或者预先计算。
struct SurfaceData
{
uint materialFeatures;
real3 baseColor;
real specularOcclusion;
float3 normalWS;
real perceptualSmoothness;
real ambientOcclusion;
real metallic;
real coatMask;
real3 specularColor;
uint diffusionProfileHash;
real subsurfaceMask;
real transmissionMask;
real thickness;
float3 tangentWS;
real anisotropy;
real iridescenceThickness;
real iridescenceMask;
real3 geomNormalWS;
real ior;
real3 transmittanceColor;
real atDistance;
real transmittanceMask;
};
struct BuiltinData
{
real opacity;
real alphaClipTreshold;
real3 bakeDiffuseLighting;
real3 backBakeDiffuseLighting;
real shadowMask0;
real shadowMask1;
real shadowMask2;
real shadowMask3;
real3 emissiveColor;
real2 motionVector;
real2 distortion;
real distortionBlur;
uint isLightmap;
uint renderingLayers;
float depthOffset;
#if defined(UNITY_VIRTUAL_TEXTURING)
real4 vtPackedFeedback;
#endif
};
看结构体就能知道这里函数的作用是什么了。
那就开看看这个函数
双面法线
前面的和LodFade相关的先略过。
先看看和双面法线相关的
float3 doubleSidedConstants = GetDoubleSidedConstants();
ApplyDoubleSidedFlipOrMirror(input, doubleSidedConstants);
对应材质中的Mirror和Flip。定义了反面法线的朝向。
这里法线存在了input.tangentToWorld[2]中。
UV
LayerTexCoord layerTexCoord;
ZERO_INITIALIZE(LayerTexCoord, layerTexCoord);
GetLayerTexCoord(input, layerTexCoord);
预处理UV。
struct LayerTexCoord
{
#ifndef LAYERED_LIT_SHADER
UVMapping base;
UVMapping details;
#else
// Regular texcoord
UVMapping base0;
UVMapping base1;
UVMapping base2;
UVMapping base3;
UVMapping details0;
UVMapping details1;
UVMapping details2;
UVMapping details3;
// Dedicated for blend mask
UVMapping blendMask;
#endif
// Store information that will be share by all UVMapping
float3 vertexNormalWS; // TODO: store also object normal map for object triplanar
float3 triplanarWeights;
#ifdef SURFACE_GRADIENT
// tangent basis for each UVSet - up to 4 for now
float3 vertexTangentWS0, vertexBitangentWS0;
float3 vertexTangentWS1, vertexBitangentWS1;
float3 vertexTangentWS2, vertexBitangentWS2;
float3 vertexTangentWS3, vertexBitangentWS3;
#endif
};
struct UVMapping
{
int mappingType;
float2 uv; // Current uv or planar uv
// Triplanar specific
float2 uvZY;
float2 uvXZ;
float2 uvXY;
float3 normalWS; // vertex normal
float3 triplanarWeights;
#ifdef SURFACE_GRADIENT
// tangent basis to use when mappingType is UV_MAPPING_UVSET
// these are vertex level in world space
float3 tangentWS;
float3 bitangentWS;
// TODO: store also object normal map for object triplanar
#endif
};
LAYERED_LIT_SHADER 是分层,先略过。
所以共有两种UVMapping,base与details。
这也就要求相关的贴图的UV分布要一致。
观察UVMapping能看出这里还考虑了三平面映射UV。
void GetLayerTexCoord(FragInputs input, inout LayerTexCoord layerTexCoord)
{
#ifdef SURFACE_GRADIENT
GenerateLayerTexCoordBasisTB(input, layerTexCoord);
#endif
GetLayerTexCoord( input.texCoord0.xy, input.texCoord1.xy, input.texCoord2.xy, input.texCoord3.xy,
input.positionRWS, input.tangentToWorld[2].xyz, layerTexCoord);
}
GenerateLayerTexCoordBasisTB用来计算切线空间
GetLayerTexCoord用来生成对应的UVMapping
高度图
#if !defined(SHADER_STAGE_RAY_TRACING)
float depthOffset = ApplyPerPixelDisplacement(input, V, layerTexCoord);
#ifdef _DEPTHOFFSET_ON
ApplyDepthOffsetPositionInput(V, depthOffset, GetViewForwardDir(), GetWorldToHClipMatrix(), posInput);
#endif
#else
float depthOffset = 0.0;
#endif
深度图相关,包含了深度偏移。
透明度裁切
#if defined(_ALPHATEST_ON)
float alphaTex = SAMPLE_UVMAPPING_TEXTURE2D(_BaseColorMap, sampler_BaseColorMap, layerTexCoord.base).a;
alphaTex = lerp(_AlphaRemapMin, _AlphaRemapMax, alphaTex);
float alphaValue = alphaTex * _BaseColor.a;
// Perform alha test very early to save performance (a killed pixel will not sample textures)
#if SHADERPASS == SHADERPASS_TRANSPARENT_DEPTH_PREPASS
float alphaCutoff = _AlphaCutoffPrepass;
#elif SHADERPASS == SHADERPASS_TRANSPARENT_DEPTH_POSTPASS
float alphaCutoff = _AlphaCutoffPostpass;
#elif (SHADERPASS == SHADERPASS_SHADOWS) || (SHADERPASS == SHADERPASS_RAYTRACING_VISIBILITY)
float alphaCutoff = _UseShadowThreshold ? _AlphaCutoffShadow : _AlphaCutoff;
#else
float alphaCutoff = _AlphaCutoff;
#endif
GENERIC_ALPHA_TEST(alphaValue, alphaCutoff);
#endif
不是重点先略过
SurfaceData
float3 normalTS;
float3 bentNormalTS;
float3 bentNormalWS;
float alpha = GetSurfaceData(input, layerTexCoord, surfaceData, normalTS, bentNormalTS);
// This need to be init here to quiet the compiler in case of decal, but can be override later.
surfaceData.geomNormalWS = input.tangentToWorld[2];
surfaceData.specularOcclusion = 1.0;
来看看GetSurfaceData这个函数
主要就是采样各种贴图,获取SurfaceData。
贴花
#if HAVE_DECALS && (defined(DECAL_SURFACE_GRADIENT) && defined(SURFACE_GRADIENT))
if (_EnableDecals)
{
DecalSurfaceData decalSurfaceData = GetDecalSurfaceData(posInput, input, alpha);
ApplyDecalToSurfaceData(decalSurfaceData, input.tangentToWorld[2], surfaceData, normalTS);
}
#endif
也先略过吧。
次法线
#ifdef _BENTNORMALMAP
GetNormalWS(input, bentNormalTS, bentNormalWS, doubleSidedConstants);
#else
bentNormalWS = surfaceData.normalWS;
#endif
获取次法线
高光遮挡
#if defined(_SPECULAR_OCCLUSION_FROM_BENT_NORMAL_MAP)
// If we have bent normal and ambient occlusion, process a specular occlusion
surfaceData.specularOcclusion = GetSpecularOcclusionFromBentAO(V, bentNormalWS, surfaceData.normalWS, surfaceData.ambientOcclusion, PerceptualSmoothnessToRoughness(surfaceData.perceptualSmoothness));
// Don't do spec occ from Ambient if there is no mask mask
#elif defined(_MASKMAP) && !defined(_SPECULAR_OCCLUSION_NONE)
surfaceData.specularOcclusion = GetSpecularOcclusionFromAmbientOcclusion(ClampNdotV(dot(surfaceData.normalWS, V)), surfaceData.ambientOcclusion, PerceptualSmoothnessToRoughness(surfaceData.perceptualSmoothness));
#endif
如果启用了次法线贴图的镜面遮挡,就修改specularOcclusion属性。
几何高光AA
#if defined(_ENABLE_GEOMETRIC_SPECULAR_AA) && !defined(SHADER_STAGE_RAY_TRACING)
// Specular AA
#ifdef PROJECTED_SPACE_NDF_FILTERING
surfaceData.perceptualSmoothness = ProjectedSpaceGeometricNormalFiltering(surfaceData.perceptualSmoothness, input.tangentToWorld[2], _SpecularAAScreenSpaceVariance, _SpecularAAThreshold);
#else
surfaceData.perceptualSmoothness = GeometricNormalFiltering(surfaceData.perceptualSmoothness, input.tangentToWorld[2], _SpecularAAScreenSpaceVariance, _SpecularAAThreshold);
#endif
#endif
用来处理几何锐角附近的高光锯齿,是通过修改光滑度来实现的,看来这一步可以预先在SP等贴图软件中提前处理,能节省一些性能。
BuiltinData
GetBuiltinData(input, V, posInput, surfaceData, alpha, bentNormalWS, depthOffset, layerTexCoord.base, builtinData);
SurfaceDatas是和表面相关的属性,和材质相关。BuiltinData则是和环境相关,如全局光、阴影、透明度裁切阙值等。
GI有三种方式,光照贴图,球谐函数,还是新退出的用3D贴图来采样的球谐函数。
PostInitBuiltinData
然后是最后的修改,
ModifyBakedDiffuseLighting(V, posInput, surfaceData, builtinData);
float multiplier = GetIndirectDiffuseMultiplier(builtinData.renderingLayers);
builtinData.bakeDiffuseLighting *= multiplier;
都是对GI的修改。
void ModifyBakedDiffuseLighting(float3 V, PositionInputs posInput, SurfaceData surfaceData, inout BuiltinData builtinData)
{
// Since this is called early at PostInitBuiltinData and we need some fields from bsdfData and preLightData,
// we get the whole structures redundantly earlier here - compiler should optimize out everything.
BSDFData bsdfData = ConvertSurfaceDataToBSDFData(posInput.positionSS, surfaceData);
PreLightData preLightData = GetPreLightData(V, posInput, bsdfData);
ModifyBakedDiffuseLighting(V, posInput, preLightData, bsdfData, builtinData);
}
又出现了两个结构体
struct BSDFData
{
uint materialFeatures;
real3 diffuseColor;
real3 fresnel0;
real ambientOcclusion;
real specularOcclusion;
float3 normalWS;
real perceptualRoughness;
real coatMask;
uint diffusionProfileIndex;
real subsurfaceMask;
real thickness;
bool useThickObjectMode;
real3 transmittance;
float3 tangentWS;
float3 bitangentWS;
real roughnessT;
real roughnessB;
real anisotropy;
real iridescenceThickness;
real iridescenceMask;
real coatRoughness;
real3 geomNormalWS;
real ior;
real3 absorptionCoefficient;
real transmittanceMask;
};
struct PreLightData
{
float NdotV; // Could be negative due to normal mapping, use ClampNdotV()
// GGX
float partLambdaV;
float energyCompensation;
// IBL
float3 iblR; // Reflected specular direction, used for IBL in EvaluateBSDF_Env()
float iblPerceptualRoughness;
float3 specularFGD; // Store preintegrated BSDF for both specular and diffuse
float diffuseFGD;
// Area lights (17 VGPRs)
// TODO: 'orthoBasisViewNormal' is just a rotation around the normal and should thus be just 1x VGPR.
float3x3 orthoBasisViewNormal; // Right-handed view-dependent orthogonal basis around the normal (6x VGPRs)
float3x3 ltcTransformDiffuse; // Inverse transformation for Lambertian or Disney Diffuse (4x VGPRs)
float3x3 ltcTransformSpecular; // Inverse transformation for GGX (4x VGPRs)
// Clear coat
float coatPartLambdaV;
float3 coatIblR;
float coatIblF; // Fresnel term for view vector
float coatReflectionWeight; // like reflectionHierarchyWeight but used to distinguish coat contribution between SSR/IBL lighting
float3x3 ltcTransformCoat; // Inverse transformation for GGX (4x VGPRs)
#if HAS_REFRACTION
// Refraction
float3 transparentRefractV; // refracted view vector after exiting the shape
float3 transparentPositionWS; // start of the refracted ray after exiting the shape
float3 transparentTransmittance; // transmittance due to absorption
float transparentSSMipLevel; // mip level of the screen space gaussian pyramid for rough refraction
#endif
};
前者是针对BSDF所需的全部参数的汇总,后者就是光源相关的属性。
来看看这两个结构体的获取
ConvertSurfaceDataToBSDFData
也是顾名思义,填充一些bsdf相关的属性。
如ao,高光,粗糙,金属,菲涅尔等等。
这里还根据材质需求的功能的不同,如是否需要SSS,曲面细分,各向异性,虹彩,清漆来填充不同的属性。
GetPreLightData
同样的,根据不同的材质,会有针对性的修改,比如为了虹彩和清漆会修改菲涅尔,
这里对于面光源,环境照明,的参数是从一张预处理的贴图中采样的:GetPreIntegratedFGDGGXAndDisneyDiffuse
还有一些奇怪的属性energyCompensation、面光源相关等等,等用到时再回过头来看吧。
LightLoopOutput
获取完大部分的属性后终于来到了光照计算
struct LightLoopOutput
{
float3 diffuseLighting;
float3 specularLighting;
};
输出很直接漫反射和高光。
LightLoopContext
struct LightLoopContext
{
int sampleReflection;
#ifdef APPLY_FOG_ON_SKY_REFLECTIONS
float3 positionWS; // For sky reflection, we need position to evalute height base fog
#endif
HDShadowContext shadowContext;
uint contactShadow; // a bit mask of 24 bits that tell if the pixel is in a contact shadow or not
real contactShadowFade; // combined fade factor of all contact shadows
SHADOW_TYPE shadowValue; // Stores the value of the cascade shadow map
real splineVisibility; // Stores the value of the cascade shadow map (unbiased for splines)
};
struct HDShadowContext
{
StructuredBuffer<HDShadowData> shadowDatas;
HDDirectionalShadowData directionalShadowData;
};
struct HDShadowData
{
float3 rot0;
float3 rot1;
float3 rot2;
float3 pos;
float4 proj;
float2 atlasOffset;
float worldTexelSize;
float normalBias;
real4 zBufferParam;
float4 shadowMapSize;
float4 shadowFilterParams0;
float3 cacheTranslationDelta;
float isInCachedAtlas;
float4x4 shadowToWorld;
};
struct HDDirectionalShadowData
{
float4 sphereCascades[4];
real4 cascadeDirection;
real cascadeBorders[4];
};
一层套一层
大部分都是定义的阴影
StructuredBuffer<HDShadowData> _HDShadowDatas;
// Only the first element is used since we only support one directional light
StructuredBuffer<HDDirectionalShadowData> _HDDirectionalShadowData;
HDShadowContext InitShadowContext()
{
HDShadowContext sc;
sc.shadowDatas = _HDShadowDatas;
sc.directionalShadowData = _HDDirectionalShadowData[0];
return sc;
}
结构体的值是整个获取的。
ApplyCameraRelativeXR(posInput.positionWS);
// Initialize the contactShadow and contactShadowFade fields
InitContactShadow(posInput, context);
适配XR中的摄像机相关世界坐标。
获取级联阴影参数。
平行光阴影
DirectionalLightData light = _DirectionalLightDatas[_DirectionalShadowIndex];
获取平行光数据
float3 L = -light.forward;
方向
然后就是采样阴影贴图。结果存到shadowValue中。
点光、附加光直接
struct AggregateLighting
{
DirectLighting direct;
IndirectLighting indirect;
};
这里其实存的是全部附加光照的总和。
遍历全部的附加光
while (v_lightListOffset < lightCount)
#endif
{
v_lightIdx = FetchIndex(lightStart, v_lightListOffset);
#if SCALARIZE_LIGHT_LOOP
uint s_lightIdx = ScalarizeElementIndex(v_lightIdx, fastPath);
#else
uint s_lightIdx = v_lightIdx;
#endif
if (s_lightIdx == -1)
break;
LightData s_lightData = FetchLight(s_lightIdx);
// If current scalar and vector light index match, we process the light. The v_lightListOffset for current thread is increased.
// Note that the following should really be ==, however, since helper lanes are not considered by WaveActiveMin, such helper lanes could
// end up with a unique v_lightIdx value that is smaller than s_lightIdx hence being stuck in a loop. All the active lanes will not have this problem.
if (s_lightIdx >= v_lightIdx)
{
v_lightListOffset++;
if (IsMatchingLightLayer(s_lightData.lightLayers, builtinData.renderingLayers))
{
DirectLighting lighting = EvaluateBSDF_Punctual(context, V, posInput, preLightData, s_lightData, bsdfData, builtinData);
AccumulateDirectLighting(lighting, aggregateLighting);
}
}
}
遍历完累加到一起。
平行光直接
uint i = 0; // Declare once to avoid the D3D11 compiler warning.
if (featureFlags & LIGHTFEATUREFLAGS_DIRECTIONAL)
{
for (i = 0; i < _DirectionalLightCount; ++i)
{
if (IsMatchingLightLayer(_DirectionalLightDatas[i].lightLayers, builtinData.renderingLayers))
{
DirectLighting lighting = EvaluateBSDF_Directional(context, V, posInput, preLightData, _DirectionalLightDatas[i], bsdfData, builtinData);
AccumulateDirectLighting(lighting, aggregateLighting);
}
}
}
循环计算全部的平行光的直接照明
面光源直接
#if SHADEROPTIONS_AREA_LIGHTS
if (featureFlags & LIGHTFEATUREFLAGS_AREA)
{
uint lightCount, lightStart;
#ifndef LIGHTLOOP_DISABLE_TILE_AND_CLUSTER
GetCountAndStart(posInput, LIGHTCATEGORY_AREA, lightStart, lightCount);
#else
lightCount = _AreaLightCount;
lightStart = _PunctualLightCount;
#endif
// COMPILER BEHAVIOR WARNING!
// If rectangle lights are before line lights, the compiler will duplicate light matrices in VGPR because they are used differently between the two types of lights.
// By keeping line lights first we avoid this behavior and save substantial register pressure.
// TODO: This is based on the current Lit.shader and can be different for any other way of implementing area lights, how to be generic and ensure performance ?
if (lightCount > 0)
{
i = 0;
uint last = lightCount - 1;
LightData lightData = FetchLight(lightStart, i);
while (i <= last && lightData.lightType == GPULIGHTTYPE_TUBE)
{
lightData.lightType = GPULIGHTTYPE_TUBE; // Enforce constant propagation
lightData.cookieMode = COOKIEMODE_NONE; // Enforce constant propagation
if (IsMatchingLightLayer(lightData.lightLayers, builtinData.renderingLayers))
{
DirectLighting lighting = EvaluateBSDF_Area(context, V, posInput, preLightData, lightData, bsdfData, builtinData);
AccumulateDirectLighting(lighting, aggregateLighting);
}
lightData = FetchLight(lightStart, min(++i, last));
}
while (i <= last) // GPULIGHTTYPE_RECTANGLE
{
lightData.lightType = GPULIGHTTYPE_RECTANGLE; // Enforce constant propagation
if (IsMatchingLightLayer(lightData.lightLayers, builtinData.renderingLayers))
{
DirectLighting lighting = EvaluateBSDF_Area(context, V, posInput, preLightData, lightData, bsdfData, builtinData);
AccumulateDirectLighting(lighting, aggregateLighting);
}
lightData = FetchLight(lightStart, min(++i, last));
}
}
}
#endif
面光源,先略过吧
直接光照
可以看出直接光照都是在类似
DirectLighting lighting = EvaluateBSDF_Directional(context, V, posInput,
preLightData, _DirectionalLightDatas[i], bsdfData, builtinData);
这之类的函数中计算的
然后累加起来,所以这里来看看这个函数
ShadeSurface_Directional
EvaluateLight_Directional
处理雾气和大气相关,修改light颜色
SHADOW_TYPE shadow = EvaluateShadow_Directional(lightLoopContext, posInput, light, builtinData, GetNormalForShadowBias(bsdfData));
float NdotL = dot(bsdfData.normalWS, L); // No microshadowing when facing away from light (use for thin transmission as well)
shadow *= NdotL >= 0.0 ? ComputeMicroShadowing(GetAmbientOcclusionForMicroShadowing(bsdfData), NdotL, _MicroShadowOpacity) : 1.0;
lightColor.rgb *= ComputeShadowColor(shadow, light.shadowTint, light.penumbraTint);
计算阴影
ShadeSurface_Infinitesimal
处理反射与投射
EvaluateBSDF
到这一步才用到了BSDF函数
总结
为了之后的理解,还是按照HDRP的文件和函数分类结构,在URP中把URP的Lit重写一遍,最终目的是在URP中使用HDRP的Lit。