400 lines
17 KiB
C#
400 lines
17 KiB
C#
using UnityEngine;
|
|
using UnityEngine.SceneManagement;
|
|
#if CINEMACHINE_POST_PROCESSING_V2
|
|
using System.Collections.Generic;
|
|
using UnityEngine.Rendering.PostProcessing;
|
|
#endif
|
|
|
|
namespace Cinemachine.PostFX
|
|
{
|
|
#if !CINEMACHINE_POST_PROCESSING_V2
|
|
// Workaround for Unity scripting bug
|
|
/// <summary>
|
|
/// This behaviour is a liaison between Cinemachine with the Post-Processing v2 module. You must
|
|
/// have the Post-Processing V2 stack package installed in order to use this behaviour.
|
|
///
|
|
/// As a component on the Virtual Camera, it holds
|
|
/// a Post-Processing Profile asset that will be applied to the Unity camera whenever
|
|
/// the Virtual camera is live. It also has the optional functionality of animating
|
|
/// the Focus Distance and DepthOfField properties of the Camera State, and
|
|
/// applying them to the current Post-Processing profile, provided that profile has a
|
|
/// DepthOfField effect that is enabled.
|
|
/// </summary>
|
|
[SaveDuringPlay]
|
|
[AddComponentMenu("")] // Hide in menu
|
|
public class CinemachinePostProcessing : CinemachineExtension
|
|
{
|
|
/// <summary>Apply PostProcessing effects</summary>
|
|
/// <param name="vcam">The virtual camera being processed</param>
|
|
/// <param name="stage">The current pipeline stage</param>
|
|
/// <param name="state">The current virtual camera state</param>
|
|
/// <param name="deltaTime">The current applicable deltaTime</param>
|
|
protected override void PostPipelineStageCallback(
|
|
CinemachineVirtualCameraBase vcam,
|
|
CinemachineCore.Stage stage, ref CameraState state, float deltaTime) {}
|
|
}
|
|
#else
|
|
/// <summary>
|
|
/// This behaviour is a liaison between Cinemachine with the Post-Processing v2 module. You must
|
|
/// have the Post-Processing V2 stack package installed in order to use this behaviour.
|
|
///
|
|
/// As a component on the Virtual Camera, it holds
|
|
/// a Post-Processing Profile asset that will be applied to the Unity camera whenever
|
|
/// the Virtual camera is live. It also has the optional functionality of animating
|
|
/// the Focus Distance and DepthOfField properties of the Camera State, and
|
|
/// applying them to the current Post-Processing profile, provided that profile has a
|
|
/// DepthOfField effect that is enabled.
|
|
/// </summary>
|
|
[DocumentationSorting(DocumentationSortingAttribute.Level.UserRef)]
|
|
[ExecuteAlways]
|
|
[AddComponentMenu("")] // Hide in menu
|
|
[SaveDuringPlay]
|
|
[DisallowMultipleComponent]
|
|
[HelpURL(Documentation.BaseURL + "manual/CinemachinePostProcessing.html")]
|
|
public class CinemachinePostProcessing : CinemachineExtension
|
|
{
|
|
/// <summary>
|
|
/// This is the priority for the vcam's PostProcessing volumes. It's set to a high
|
|
/// number in order to ensure that it overrides other volumes for the active vcam.
|
|
/// You can change this value if necessary to work with other systems.
|
|
/// </summary>
|
|
static public float s_VolumePriority = 1000f;
|
|
|
|
/// <summary>This is obsolete, please use m_FocusTracking</summary>
|
|
[HideInInspector]
|
|
public bool m_FocusTracksTarget;
|
|
|
|
/// <summary>The reference object for focus tracking</summary>
|
|
public enum FocusTrackingMode
|
|
{
|
|
/// <summary>No focus tracking</summary>
|
|
None,
|
|
/// <summary>Focus offset is relative to the LookAt target</summary>
|
|
LookAtTarget,
|
|
/// <summary>Focus offset is relative to the Follow target</summary>
|
|
FollowTarget,
|
|
/// <summary>Focus offset is relative to the Custom target set here</summary>
|
|
CustomTarget,
|
|
/// <summary>Focus offset is relative to the camera</summary>
|
|
Camera
|
|
};
|
|
|
|
/// <summary>If the profile has the appropriate overrides, will set the base focus
|
|
/// distance to be the distance from the selected target to the camera.
|
|
/// The Focus Offset field will then modify that distance</summary>
|
|
[Tooltip("If the profile has the appropriate overrides, will set the base focus "
|
|
+ "distance to be the distance from the selected target to the camera."
|
|
+ "The Focus Offset field will then modify that distance.")]
|
|
public FocusTrackingMode m_FocusTracking;
|
|
|
|
/// <summary>The target to use if Focus Tracks Target is set to Custom Target</summary>
|
|
[Tooltip("The target to use if Focus Tracks Target is set to Custom Target")]
|
|
public Transform m_FocusTarget;
|
|
|
|
/// <summary>Offset from target distance, to be used with Focus Tracks Target.
|
|
/// Offsets the sharpest point away from the location of the focus target</summary>
|
|
[Tooltip("Offset from target distance, to be used with Focus Tracks Target. "
|
|
+ "Offsets the sharpest point away from the location of the focus target.")]
|
|
public float m_FocusOffset;
|
|
|
|
/// <summary>
|
|
/// This Post-Processing profile will be applied whenever this virtual camera is live
|
|
/// </summary>
|
|
[Tooltip("This Post-Processing profile will be applied whenever this virtual camera is live")]
|
|
public PostProcessProfile m_Profile;
|
|
|
|
class VcamExtraState
|
|
{
|
|
public PostProcessProfile mProfileCopy;
|
|
|
|
public void CreateProfileCopy(PostProcessProfile source)
|
|
{
|
|
DestroyProfileCopy();
|
|
PostProcessProfile profile = ScriptableObject.CreateInstance<PostProcessProfile>();
|
|
if (source != null)
|
|
{
|
|
foreach (var item in source.settings)
|
|
{
|
|
var itemCopy = Instantiate(item);
|
|
profile.settings.Add(itemCopy);
|
|
}
|
|
}
|
|
mProfileCopy = profile;
|
|
}
|
|
|
|
public void DestroyProfileCopy()
|
|
{
|
|
if (mProfileCopy != null)
|
|
RuntimeUtility.DestroyObject(mProfileCopy);
|
|
mProfileCopy = null;
|
|
}
|
|
}
|
|
|
|
/// <summary>True if the profile is enabled and nontrivial</summary>
|
|
public bool IsValid { get { return m_Profile != null && m_Profile.settings.Count > 0; } }
|
|
|
|
/// <summary>Called by the editor when the shared asset has been edited</summary>
|
|
public void InvalidateCachedProfile()
|
|
{
|
|
var list = GetAllExtraStates<VcamExtraState>();
|
|
for (int i = 0; i < list.Count; ++i)
|
|
list[i].DestroyProfileCopy();
|
|
}
|
|
|
|
protected override void OnEnable()
|
|
{
|
|
base.OnEnable();
|
|
|
|
// Map legacy m_FocusTracksTarget to focus mode
|
|
if (m_FocusTracksTarget)
|
|
{
|
|
m_FocusTracking = VirtualCamera.LookAt != null
|
|
? FocusTrackingMode.LookAtTarget : FocusTrackingMode.Camera;
|
|
}
|
|
m_FocusTracksTarget = false;
|
|
}
|
|
|
|
protected override void OnDestroy()
|
|
{
|
|
InvalidateCachedProfile();
|
|
base.OnDestroy();
|
|
}
|
|
|
|
/// <summary>Apply PostProcessing effects</summary>
|
|
/// <param name="vcam">The virtual camera being processed</param>
|
|
/// <param name="stage">The current pipeline stage</param>
|
|
/// <param name="state">The current virtual camera state</param>
|
|
/// <param name="deltaTime">The current applicable deltaTime</param>
|
|
protected override void PostPipelineStageCallback(
|
|
CinemachineVirtualCameraBase vcam,
|
|
CinemachineCore.Stage stage, ref CameraState state, float deltaTime)
|
|
{
|
|
// Set the focus after the camera has been fully positioned.
|
|
if (stage == CinemachineCore.Stage.Finalize)
|
|
{
|
|
var extra = GetExtraState<VcamExtraState>(vcam);
|
|
if (!IsValid)
|
|
extra.DestroyProfileCopy();
|
|
else
|
|
{
|
|
var profile = m_Profile;
|
|
|
|
// Handle Follow Focus
|
|
if (m_FocusTracking == FocusTrackingMode.None)
|
|
extra.DestroyProfileCopy();
|
|
else
|
|
{
|
|
if (extra.mProfileCopy == null)
|
|
extra.CreateProfileCopy(m_Profile);
|
|
profile = extra.mProfileCopy;
|
|
DepthOfField dof;
|
|
if (profile.TryGetSettings(out dof))
|
|
{
|
|
float focusDistance = m_FocusOffset;
|
|
if (m_FocusTracking == FocusTrackingMode.LookAtTarget)
|
|
focusDistance += (state.FinalPosition - state.ReferenceLookAt).magnitude;
|
|
else
|
|
{
|
|
Transform focusTarget = null;
|
|
switch (m_FocusTracking)
|
|
{
|
|
default: break;
|
|
case FocusTrackingMode.FollowTarget: focusTarget = VirtualCamera.Follow; break;
|
|
case FocusTrackingMode.CustomTarget: focusTarget = m_FocusTarget; break;
|
|
}
|
|
if (focusTarget != null)
|
|
focusDistance += (state.FinalPosition - focusTarget.position).magnitude;
|
|
}
|
|
#if UNITY_2022_2_OR_NEWER
|
|
state.Lens.FocusDistance =
|
|
#endif
|
|
dof.focusDistance.value = Mathf.Max(0, focusDistance);
|
|
}
|
|
}
|
|
|
|
// Apply the post-processing
|
|
state.AddCustomBlendable(new CameraState.CustomBlendable(profile, 1));
|
|
}
|
|
}
|
|
}
|
|
|
|
static void OnCameraCut(CinemachineBrain brain)
|
|
{
|
|
// Debug.Log("Camera cut event");
|
|
PostProcessLayer postFX = GetPPLayer(brain);
|
|
if (postFX != null)
|
|
postFX.ResetHistory();
|
|
}
|
|
|
|
static void ApplyPostFX(CinemachineBrain brain)
|
|
{
|
|
PostProcessLayer ppLayer = GetPPLayer(brain);
|
|
if (ppLayer == null || !ppLayer.enabled || ppLayer.volumeLayer == 0)
|
|
return;
|
|
|
|
CameraState state = brain.CurrentCameraState;
|
|
int numBlendables = state.NumCustomBlendables;
|
|
List<PostProcessVolume> volumes = GetDynamicBrainVolumes(brain, ppLayer, numBlendables);
|
|
for (int i = 0; i < volumes.Count; ++i)
|
|
{
|
|
volumes[i].weight = 0;
|
|
volumes[i].sharedProfile = null;
|
|
volumes[i].profile = null;
|
|
}
|
|
PostProcessVolume firstVolume = null;
|
|
int numPPblendables = 0;
|
|
for (int i = 0; i < numBlendables; ++i)
|
|
{
|
|
var b = state.GetCustomBlendable(i);
|
|
var profile = b.m_Custom as PostProcessProfile;
|
|
if (!(profile == null)) // in case it was deleted
|
|
{
|
|
PostProcessVolume v = volumes[i];
|
|
if (firstVolume == null)
|
|
firstVolume = v;
|
|
v.sharedProfile = profile;
|
|
v.isGlobal = true;
|
|
v.priority = s_VolumePriority - (numBlendables - i) - 1;
|
|
v.weight = b.m_Weight;
|
|
++numPPblendables;
|
|
}
|
|
#if false // set this to true to force first weight to 1
|
|
// If more than one volume, then set the frst one's weight to 1
|
|
if (numPPblendables > 1)
|
|
firstVolume.weight = 1;
|
|
#endif
|
|
}
|
|
}
|
|
|
|
static string sVolumeOwnerName = "__CMVolumes";
|
|
static List<PostProcessVolume> sVolumes = new List<PostProcessVolume>();
|
|
static List<PostProcessVolume> GetDynamicBrainVolumes(
|
|
CinemachineBrain brain, PostProcessLayer ppLayer, int minVolumes)
|
|
{
|
|
// Locate the camera's child object that holds our dynamic volumes
|
|
GameObject volumeOwner = null;
|
|
Transform t = brain.transform;
|
|
int numChildren = t.childCount;
|
|
|
|
sVolumes.Clear();
|
|
for (int i = 0; volumeOwner == null && i < numChildren; ++i)
|
|
{
|
|
GameObject child = t.GetChild(i).gameObject;
|
|
if (child.hideFlags == HideFlags.HideAndDontSave)
|
|
{
|
|
child.GetComponents(sVolumes);
|
|
if (sVolumes.Count > 0)
|
|
volumeOwner = child;
|
|
}
|
|
}
|
|
|
|
if (minVolumes > 0)
|
|
{
|
|
if (volumeOwner == null)
|
|
{
|
|
volumeOwner = new GameObject(sVolumeOwnerName);
|
|
volumeOwner.hideFlags = HideFlags.HideAndDontSave;
|
|
volumeOwner.transform.parent = t;
|
|
}
|
|
// Update the volume's layer so it will be seen
|
|
int mask = ppLayer.volumeLayer.value;
|
|
for (int i = 0; i < 32; ++i)
|
|
{
|
|
if ((mask & (1 << i)) != 0)
|
|
{
|
|
volumeOwner.layer = i;
|
|
break;
|
|
}
|
|
}
|
|
while (sVolumes.Count < minVolumes)
|
|
sVolumes.Add(volumeOwner.gameObject.AddComponent<PostProcessVolume>());
|
|
}
|
|
return sVolumes;
|
|
}
|
|
|
|
static Dictionary<CinemachineBrain, PostProcessLayer> mBrainToLayer
|
|
= new Dictionary<CinemachineBrain, PostProcessLayer>();
|
|
|
|
static PostProcessLayer GetPPLayer(CinemachineBrain brain)
|
|
{
|
|
bool found = mBrainToLayer.TryGetValue(brain, out PostProcessLayer layer);
|
|
if (layer != null)
|
|
return layer; // layer is valid and in our lookup
|
|
|
|
// If the layer in the lookup table is a deleted object, we must remove
|
|
// the brain's callback for it
|
|
if (found && !ReferenceEquals(layer, null))
|
|
{
|
|
// layer is a deleted object
|
|
brain.m_CameraCutEvent.RemoveListener(OnCameraCut);
|
|
mBrainToLayer.Remove(brain);
|
|
layer = null;
|
|
found = false;
|
|
}
|
|
|
|
// Brain is not in our lookup - add it.
|
|
#if UNITY_2019_2_OR_NEWER
|
|
brain.TryGetComponent(out layer);
|
|
if (layer != null)
|
|
{
|
|
brain.m_CameraCutEvent.AddListener(OnCameraCut); // valid layer
|
|
mBrainToLayer[brain] = layer;
|
|
}
|
|
#else
|
|
// In order to avoid calling GetComponent() every frame in the case
|
|
// where there is legitimately no layer on the brain, we will add
|
|
// null to the lookup table if no layer is present.
|
|
if (!found)
|
|
{
|
|
layer = brain.GetComponent<PostProcessLayer>();
|
|
if (layer != null)
|
|
brain.m_CameraCutEvent.AddListener(OnCameraCut); // valid layer
|
|
|
|
// Exception: never add null in the case where user adds a layer while
|
|
// in the editor. If we were to add null in this case, then the new
|
|
// layer would not be detected. We are willing to live with
|
|
// calling GetComponent() every frame while in edit mode.
|
|
if (Application.isPlaying || layer != null)
|
|
mBrainToLayer[brain] = layer;
|
|
}
|
|
#endif
|
|
return layer;
|
|
}
|
|
|
|
static void CleanupLookupTable()
|
|
{
|
|
var iter = mBrainToLayer.GetEnumerator();
|
|
while (iter.MoveNext())
|
|
{
|
|
var brain = iter.Current.Key;
|
|
if (brain != null)
|
|
brain.m_CameraCutEvent.RemoveListener(OnCameraCut);
|
|
}
|
|
mBrainToLayer.Clear();
|
|
}
|
|
|
|
#if UNITY_EDITOR
|
|
[UnityEditor.InitializeOnLoad]
|
|
class EditorInitialize
|
|
{
|
|
static EditorInitialize()
|
|
{
|
|
UnityEditor.EditorApplication.playModeStateChanged += (pmsc) => CleanupLookupTable();
|
|
InitializeModule();
|
|
}
|
|
}
|
|
#endif
|
|
[RuntimeInitializeOnLoadMethod]
|
|
static void InitializeModule()
|
|
{
|
|
// After the brain pushes the state to the camera, hook in to the PostFX
|
|
CinemachineCore.CameraUpdatedEvent.RemoveListener(ApplyPostFX);
|
|
CinemachineCore.CameraUpdatedEvent.AddListener(ApplyPostFX);
|
|
|
|
// Clean up our resources
|
|
SceneManager.sceneUnloaded += (scene) => CleanupLookupTable();
|
|
}
|
|
}
|
|
#endif
|
|
}
|