summaryrefslogtreecommitdiffstats
path: root/src/video_core/texture_cache
diff options
context:
space:
mode:
authorFernando Sahmkow <fsahmkow27@gmail.com>2019-05-07 16:57:16 +0200
committerReinUsesLisp <reinuseslisp@airmail.cc>2019-06-21 02:36:12 +0200
commitbc930754cc9437ddd86e7d246b3eb4302540896a (patch)
tree75e56e5e041db7b327a13c25ab13bc524458dc29 /src/video_core/texture_cache
parentCorrect Surface Base and Views for new Texture Cache (diff)
downloadyuzu-bc930754cc9437ddd86e7d246b3eb4302540896a.tar
yuzu-bc930754cc9437ddd86e7d246b3eb4302540896a.tar.gz
yuzu-bc930754cc9437ddd86e7d246b3eb4302540896a.tar.bz2
yuzu-bc930754cc9437ddd86e7d246b3eb4302540896a.tar.lz
yuzu-bc930754cc9437ddd86e7d246b3eb4302540896a.tar.xz
yuzu-bc930754cc9437ddd86e7d246b3eb4302540896a.tar.zst
yuzu-bc930754cc9437ddd86e7d246b3eb4302540896a.zip
Diffstat (limited to 'src/video_core/texture_cache')
-rw-r--r--src/video_core/texture_cache/texture_cache.h454
1 files changed, 366 insertions, 88 deletions
diff --git a/src/video_core/texture_cache/texture_cache.h b/src/video_core/texture_cache/texture_cache.h
index c5c01957a..eb0d9bc10 100644
--- a/src/video_core/texture_cache/texture_cache.h
+++ b/src/video_core/texture_cache/texture_cache.h
@@ -22,6 +22,7 @@
#include "video_core/memory_manager.h"
#include "video_core/rasterizer_interface.h"
#include "video_core/surface.h"
+#include "video_core/texture_cache/copy_params.h"
#include "video_core/texture_cache/surface_base.h"
#include "video_core/texture_cache/surface_params.h"
#include "video_core/texture_cache/surface_view.h"
@@ -40,32 +41,42 @@ class RasterizerInterface;
namespace VideoCommon {
+using VideoCore::Surface::SurfaceTarget;
+using RenderTargetConfig = Tegra::Engines::Maxwell3D::Regs::RenderTargetConfig;
+
template <typename TSurface, typename TView>
class TextureCache {
- using IntervalMap = boost::icl::interval_map<CacheAddr, std::set<std::shared_ptr<TSurface>>>;
+ using IntervalMap = boost::icl::interval_map<CacheAddr, std::set<TSurface>>;
using IntervalType = typename IntervalMap::interval_type;
public:
+ void InitMemoryMananger(Tegra::MemoryManager& memory_manager) {
+ this->memory_manager = &memory_manager;
+ }
+
void InvalidateRegion(CacheAddr addr, std::size_t size) {
for (const auto& surface : GetSurfacesInRegion(addr, size)) {
- if (!surface->IsRegistered()) {
- // Skip duplicates
- continue;
- }
Unregister(surface);
}
}
- TView* GetTextureSurface(const Tegra::Texture::FullTextureInfo& config) {
+ void InvalidateRegionEx(GPUVAddr addr, std::size_t size) {
+ for (const auto& surface : GetSurfacesInRegionInner(addr, size)) {
+ Unregister(surface);
+ }
+ }
+
+ TView GetTextureSurface(const Tegra::Texture::FullTextureInfo& config,
+ const VideoCommon::Shader::Sampler& entry) {
const auto gpu_addr{config.tic.Address()};
if (!gpu_addr) {
return {};
}
- const auto params{SurfaceParams::CreateForTexture(system, config)};
- return GetSurfaceView(gpu_addr, params, true);
+ const auto params{SurfaceParams::CreateForTexture(system, config, entry)};
+ return GetSurface(gpu_addr, params, true).second;
}
- TView* GetDepthBufferSurface(bool preserve_contents) {
+ TView GetDepthBufferSurface(bool preserve_contents) {
const auto& regs{system.GPU().Maxwell3D().regs};
const auto gpu_addr{regs.zeta.Address()};
if (!gpu_addr || !regs.zeta_enable) {
@@ -75,36 +86,75 @@ public:
system, regs.zeta_width, regs.zeta_height, regs.zeta.format,
regs.zeta.memory_layout.block_width, regs.zeta.memory_layout.block_height,
regs.zeta.memory_layout.block_depth, regs.zeta.memory_layout.type)};
- return GetSurfaceView(gpu_addr, depth_params, preserve_contents);
+ auto surface_view = GetSurface(gpu_addr, depth_params, preserve_contents);
+ if (depth_buffer.target)
+ depth_buffer.target->MarkAsProtected(false);
+ if (depth_buffer.target)
+ depth_buffer.target->MarkAsProtected(true);
+ return surface_view.second;
}
- TView* GetColorBufferSurface(std::size_t index, bool preserve_contents) {
+ TView GetColorBufferSurface(std::size_t index, bool preserve_contents) {
ASSERT(index < Tegra::Engines::Maxwell3D::Regs::NumRenderTargets);
const auto& regs{system.GPU().Maxwell3D().regs};
if (index >= regs.rt_control.count || regs.rt[index].Address() == 0 ||
regs.rt[index].format == Tegra::RenderTargetFormat::NONE) {
+ SetEmptyColorBuffer(index);
return {};
}
- auto& memory_manager{system.GPU().MemoryManager()};
- const auto& config{system.GPU().Maxwell3D().regs.rt[index]};
- const auto gpu_addr{config.Address() +
- config.base_layer * config.layer_stride * sizeof(u32)};
+ const auto& config{regs.rt[index]};
+ const auto gpu_addr{config.Address()};
if (!gpu_addr) {
+ SetEmptyColorBuffer(index);
return {};
}
- return GetSurfaceView(gpu_addr, SurfaceParams::CreateForFramebuffer(system, index),
- preserve_contents);
+ auto surface_view = GetSurface(gpu_addr, SurfaceParams::CreateForFramebuffer(system, index),
+ preserve_contents);
+ if (render_targets[index].target)
+ render_targets[index].target->MarkAsProtected(false);
+ render_targets[index].target = surface_view.first;
+ if (render_targets[index].target)
+ render_targets[index].target->MarkAsProtected(true);
+ return surface_view.second;
+ }
+
+ void MarkColorBufferInUse(std::size_t index) {
+ if (render_targets[index].target)
+ render_targets[index].target->MarkAsModified(true, Tick());
}
- TView* GetFermiSurface(const Tegra::Engines::Fermi2D::Regs::Surface& config) {
- return GetSurfaceView(config.Address(), SurfaceParams::CreateForFermiCopySurface(config),
- true);
+ void MarkDepthBufferInUse() {
+ if (depth_buffer.target)
+ depth_buffer.target->MarkAsModified(true, Tick());
}
- std::shared_ptr<TSurface> TryFindFramebufferSurface(const u8* host_ptr) const {
+ void SetEmptyDepthBuffer() {
+ if (depth_buffer.target != nullptr) {
+ depth_buffer.target->MarkAsProtected(false);
+ depth_buffer.target = nullptr;
+ depth_buffer.view = nullptr;
+ }
+ }
+
+ void SetEmptyColorBuffer(std::size_t index) {
+ if (render_targets[index].target != nullptr) {
+ render_targets[index].target->MarkAsProtected(false);
+ std::memset(&render_targets[index].config, sizeof(RenderTargetConfig), 0);
+ render_targets[index].target = nullptr;
+ render_targets[index].view = nullptr;
+ }
+ }
+
+ TView GetFermiSurface(const Tegra::Engines::Fermi2D::Regs::Surface& config) {
+ SurfaceParams params = SurfaceParams::CreateForFermiCopySurface(config);
+ const GPUVAddr gpu_addr = config.Address();
+ return GetSurface(gpu_addr, params, true).second;
+ }
+
+ TSurface TryFindFramebufferSurface(const u8* host_ptr) const {
const auto it{registered_surfaces.find(ToCacheAddr(host_ptr))};
return it != registered_surfaces.end() ? *it->second.begin() : nullptr;
}
@@ -115,126 +165,334 @@ public:
protected:
TextureCache(Core::System& system, VideoCore::RasterizerInterface& rasterizer)
- : system{system}, rasterizer{rasterizer} {}
+ : system{system}, rasterizer{rasterizer} {
+ for (std::size_t i = 0; i < Tegra::Engines::Maxwell3D::Regs::NumRenderTargets; i++) {
+ SetEmptyColorBuffer(i);
+ }
+ SetEmptyDepthBuffer();
+ }
~TextureCache() = default;
- virtual TView* TryFastGetSurfaceView(
- GPUVAddr gpu_addr, VAddr cpu_addr, u8* host_ptr, const SurfaceParams& params,
- bool preserve_contents, const std::vector<std::shared_ptr<TSurface>>& overlaps) = 0;
+ virtual TSurface CreateSurface(GPUVAddr gpu_addr, const SurfaceParams& params) = 0;
- virtual std::shared_ptr<TSurface> CreateSurface(const SurfaceParams& params) = 0;
+ virtual void ImageCopy(TSurface src_surface, TSurface dst_surface,
+ const CopyParams& copy_params) = 0;
- void Register(std::shared_ptr<TSurface> surface, GPUVAddr gpu_addr, VAddr cpu_addr,
- u8* host_ptr) {
- surface->Register(gpu_addr, cpu_addr, host_ptr);
- registered_surfaces.add({GetSurfaceInterval(surface), {surface}});
- rasterizer.UpdatePagesCachedCount(surface->GetCpuAddr(), surface->GetSizeInBytes(), 1);
+ void Register(TSurface surface) {
+ const GPUVAddr gpu_addr = surface->GetGpuAddr();
+ u8* host_ptr = memory_manager->GetPointer(gpu_addr);
+ const std::size_t size = surface->GetSizeInBytes();
+ const std::optional<VAddr> cpu_addr = memory_manager->GpuToCpuAddress(gpu_addr);
+ if (!host_ptr || !cpu_addr) {
+ LOG_CRITICAL(HW_GPU, "Failed to register surface with unmapped gpu_address 0x{:016x}",
+ gpu_addr);
+ return;
+ }
+ surface->SetHostPtr(host_ptr);
+ surface->SetCpuAddr(*cpu_addr);
+ registered_surfaces.add({GetInterval(host_ptr, size), {surface}});
+ rasterizer.UpdatePagesCachedCount(*cpu_addr, size, 1);
+ RegisterInnerCache(surface);
+ surface->MarkAsRegistered(true);
}
- void Unregister(std::shared_ptr<TSurface> surface) {
- registered_surfaces.subtract({GetSurfaceInterval(surface), {surface}});
- rasterizer.UpdatePagesCachedCount(surface->GetCpuAddr(), surface->GetSizeInBytes(), -1);
- surface->Unregister();
+ void Unregister(TSurface surface) {
+ if (surface->IsProtected())
+ return;
+ const GPUVAddr gpu_addr = surface->GetGpuAddr();
+ const void* host_ptr = surface->GetHostPtr();
+ const std::size_t size = surface->GetSizeInBytes();
+ const VAddr cpu_addr = surface->GetCpuAddr();
+ registered_surfaces.erase(GetInterval(host_ptr, size));
+ rasterizer.UpdatePagesCachedCount(cpu_addr, size, -1);
+ UnregisterInnerCache(surface);
+ surface->MarkAsRegistered(false);
+ ReserveSurface(surface->GetSurfaceParams(), surface);
}
- std::shared_ptr<TSurface> GetUncachedSurface(const SurfaceParams& params) {
- if (const auto surface = TryGetReservedSurface(params); surface)
+ TSurface GetUncachedSurface(const GPUVAddr gpu_addr, const SurfaceParams& params) {
+ if (const auto surface = TryGetReservedSurface(params); surface) {
+ surface->SetGpuAddr(gpu_addr);
return surface;
+ }
// No reserved surface available, create a new one and reserve it
- auto new_surface{CreateSurface(params)};
- ReserveSurface(params, new_surface);
+ auto new_surface{CreateSurface(gpu_addr, params)};
return new_surface;
}
Core::System& system;
private:
- TView* GetSurfaceView(GPUVAddr gpu_addr, const SurfaceParams& params, bool preserve_contents) {
- auto& memory_manager{system.GPU().MemoryManager()};
- const auto cpu_addr{memory_manager.GpuToCpuAddress(gpu_addr)};
- DEBUG_ASSERT(cpu_addr);
-
- const auto host_ptr{memory_manager.GetPointer(gpu_addr)};
- const auto cache_addr{ToCacheAddr(host_ptr)};
- auto overlaps{GetSurfacesInRegion(cache_addr, params.GetGuestSizeInBytes())};
- if (overlaps.empty()) {
- return LoadSurfaceView(gpu_addr, *cpu_addr, host_ptr, params, preserve_contents);
+ enum class RecycleStrategy : u32 {
+ Ignore = 0,
+ Flush = 1,
+ BufferCopy = 3,
+ };
+
+ RecycleStrategy PickStrategy(std::vector<TSurface>& overlaps, const SurfaceParams& params,
+ const GPUVAddr gpu_addr, const bool untopological) {
+ // Untopological decision
+ if (untopological) {
+ return RecycleStrategy::Ignore;
+ }
+ // 3D Textures decision
+ if (params.block_depth > 1 || params.target == SurfaceTarget::Texture3D) {
+ return RecycleStrategy::Flush;
}
+ for (auto s : overlaps) {
+ const auto& s_params = s->GetSurfaceParams();
+ if (s_params.block_depth > 1 || s_params.target == SurfaceTarget::Texture3D) {
+ return RecycleStrategy::Flush;
+ }
+ }
+ return RecycleStrategy::Ignore;
+ }
- if (overlaps.size() == 1) {
- if (TView* view = overlaps[0]->TryGetView(gpu_addr, params); view) {
- return view;
+ std::pair<TSurface, TView> RecycleSurface(std::vector<TSurface>& overlaps,
+ const SurfaceParams& params, const GPUVAddr gpu_addr,
+ const u8* host_ptr, const bool preserve_contents,
+ const bool untopological) {
+ for (auto surface : overlaps) {
+ Unregister(surface);
+ }
+ RecycleStrategy strategy = !Settings::values.use_accurate_gpu_emulation
+ ? PickStrategy(overlaps, params, gpu_addr, untopological)
+ : RecycleStrategy::Flush;
+ switch (strategy) {
+ case RecycleStrategy::Ignore: {
+ return InitializeSurface(gpu_addr, params, preserve_contents);
+ }
+ case RecycleStrategy::Flush: {
+ std::sort(overlaps.begin(), overlaps.end(),
+ [](const TSurface& a, const TSurface& b) -> bool {
+ return a->GetModificationTick() < b->GetModificationTick();
+ });
+ for (auto surface : overlaps) {
+ FlushSurface(surface);
}
+ return InitializeSurface(gpu_addr, params, preserve_contents);
}
+ default: {
+ UNIMPLEMENTED_MSG("Unimplemented Texture Cache Recycling Strategy!");
+ return InitializeSurface(gpu_addr, params, preserve_contents);
+ }
+ }
+ }
- const auto fast_view{TryFastGetSurfaceView(gpu_addr, *cpu_addr, host_ptr, params,
- preserve_contents, overlaps)};
+ std::pair<TSurface, TView> RebuildMirage(TSurface current_surface,
+ const SurfaceParams& params) {
+ const auto gpu_addr = current_surface->GetGpuAddr();
+ TSurface new_surface = GetUncachedSurface(gpu_addr, params);
+ std::vector<CopyParams> bricks = current_surface->BreakDown();
+ for (auto& brick : bricks) {
+ ImageCopy(current_surface, new_surface, brick);
+ }
+ Unregister(current_surface);
+ Register(new_surface);
+ return {new_surface, new_surface->GetMainView()};
+ }
- if (!fast_view) {
- std::sort(overlaps.begin(), overlaps.end(), [](const auto& lhs, const auto& rhs) {
- return lhs->GetModificationTick() < rhs->GetModificationTick();
- });
+ std::pair<TSurface, TView> ManageStructuralMatch(TSurface current_surface,
+ const SurfaceParams& params) {
+ const bool is_mirage = !current_surface->MatchFormat(params.pixel_format);
+ if (is_mirage) {
+ return RebuildMirage(current_surface, params);
}
+ const bool matches_target = current_surface->MatchTarget(params.target);
+ if (matches_target) {
+ return {current_surface, current_surface->GetMainView()};
+ }
+ return {current_surface, current_surface->EmplaceOverview(params)};
+ }
- for (const auto& surface : overlaps) {
- if (!fast_view) {
- // Flush even when we don't care about the contents, to preserve memory not
- // written by the new surface.
- FlushSurface(surface);
+ std::optional<std::pair<TSurface, TView>> ReconstructSurface(std::vector<TSurface>& overlaps,
+ const SurfaceParams& params,
+ const GPUVAddr gpu_addr,
+ const u8* host_ptr) {
+ if (!params.is_layered || params.target == SurfaceTarget::Texture3D) {
+ return {};
+ }
+ TSurface new_surface = GetUncachedSurface(gpu_addr, params);
+ for (auto surface : overlaps) {
+ const SurfaceParams& src_params = surface->GetSurfaceParams();
+ if (src_params.is_layered || src_params.num_levels > 1) {
+ // We send this cases to recycle as they are more complex to handle
+ return {};
+ }
+ const std::size_t candidate_size = src_params.GetGuestSizeInBytes();
+ auto mipmap_layer = new_surface->GetLayerMipmap(surface->GetGpuAddr());
+ if (!mipmap_layer) {
+ return {};
}
+ const u32 layer = (*mipmap_layer).first;
+ const u32 mipmap = (*mipmap_layer).second;
+ if (new_surface->GetMipmapSize(mipmap) != candidate_size) {
+ return {};
+ }
+ // Now we got all the data set up
+ CopyParams copy_params{};
+ const u32 dst_width = params.GetMipWidth(mipmap);
+ const u32 dst_height = params.GetMipHeight(mipmap);
+ copy_params.width = std::min(src_params.width, dst_width);
+ copy_params.height = std::min(src_params.height, dst_height);
+ copy_params.depth = 1;
+ copy_params.source_level = 0;
+ copy_params.dest_level = mipmap;
+ copy_params.source_z = 0;
+ copy_params.dest_z = layer;
+ ImageCopy(surface, new_surface, copy_params);
+ }
+ for (auto surface : overlaps) {
Unregister(surface);
}
- if (fast_view) {
- return fast_view;
+ Register(new_surface);
+ return {{new_surface, new_surface->GetMainView()}};
+ }
+
+ std::pair<TSurface, TView> GetSurface(const GPUVAddr gpu_addr, const SurfaceParams& params,
+ bool preserve_contents) {
+
+ const auto host_ptr{memory_manager->GetPointer(gpu_addr)};
+ const auto cache_addr{ToCacheAddr(host_ptr)};
+ const std::size_t candidate_size = params.GetGuestSizeInBytes();
+ auto overlaps{GetSurfacesInRegionInner(gpu_addr, candidate_size)};
+ if (overlaps.empty()) {
+ return InitializeSurface(gpu_addr, params, preserve_contents);
+ }
+
+ for (auto surface : overlaps) {
+ if (!surface->MatchesTopology(params)) {
+ return RecycleSurface(overlaps, params, gpu_addr, host_ptr, preserve_contents,
+ true);
+ }
}
- return LoadSurfaceView(gpu_addr, *cpu_addr, host_ptr, params, preserve_contents);
+ if (overlaps.size() == 1) {
+ TSurface current_surface = overlaps[0];
+ if (current_surface->MatchesStructure(params) &&
+ current_surface->GetGpuAddr() == gpu_addr &&
+ (params.target != SurfaceTarget::Texture3D ||
+ current_surface->MatchTarget(params.target))) {
+ return ManageStructuralMatch(current_surface, params);
+ }
+ if (current_surface->GetSizeInBytes() <= candidate_size) {
+ return RecycleSurface(overlaps, params, gpu_addr, host_ptr, preserve_contents,
+ false);
+ }
+ std::optional<TView> view = current_surface->EmplaceView(params, gpu_addr);
+ if (view.has_value()) {
+ const bool is_mirage = !current_surface->MatchFormat(params.pixel_format);
+ if (is_mirage) {
+ LOG_CRITICAL(HW_GPU, "Mirage View Unsupported");
+ return RecycleSurface(overlaps, params, gpu_addr, host_ptr, preserve_contents,
+ false);
+ }
+ return {current_surface, *view};
+ }
+ return RecycleSurface(overlaps, params, gpu_addr, host_ptr, preserve_contents, false);
+ } else {
+ std::optional<std::pair<TSurface, TView>> view =
+ ReconstructSurface(overlaps, params, gpu_addr, host_ptr);
+ if (view.has_value()) {
+ return *view;
+ }
+ return RecycleSurface(overlaps, params, gpu_addr, host_ptr, preserve_contents, false);
+ }
}
- TView* LoadSurfaceView(GPUVAddr gpu_addr, VAddr cpu_addr, u8* host_ptr,
- const SurfaceParams& params, bool preserve_contents) {
- const auto new_surface{GetUncachedSurface(params)};
- Register(new_surface, gpu_addr, cpu_addr, host_ptr);
+ std::pair<TSurface, TView> InitializeSurface(GPUVAddr gpu_addr, const SurfaceParams& params,
+ bool preserve_contents) {
+ auto new_surface{GetUncachedSurface(gpu_addr, params)};
+ Register(new_surface);
if (preserve_contents) {
LoadSurface(new_surface);
}
- return new_surface->GetView(gpu_addr, params);
+ return {new_surface, new_surface->GetMainView()};
}
- void LoadSurface(const std::shared_ptr<TSurface>& surface) {
- surface->LoadBuffer();
- surface->UploadTexture();
- surface->MarkAsModified(false);
+ void LoadSurface(const TSurface& surface) {
+ staging_buffer.resize(surface->GetHostSizeInBytes());
+ surface->LoadBuffer(*memory_manager, staging_buffer);
+ surface->UploadTexture(staging_buffer);
+ surface->MarkAsModified(false, Tick());
}
- void FlushSurface(const std::shared_ptr<TSurface>& surface) {
+ void FlushSurface(const TSurface& surface) {
if (!surface->IsModified()) {
return;
}
- surface->DownloadTexture();
- surface->FlushBuffer();
+ staging_buffer.resize(surface->GetHostSizeInBytes());
+ surface->DownloadTexture(staging_buffer);
+ surface->FlushBuffer(staging_buffer);
+ surface->MarkAsModified(false, Tick());
}
- std::vector<std::shared_ptr<TSurface>> GetSurfacesInRegion(CacheAddr cache_addr,
- std::size_t size) const {
+ std::vector<TSurface> GetSurfacesInRegion(CacheAddr cache_addr, std::size_t size) const {
if (size == 0) {
return {};
}
const IntervalType interval{cache_addr, cache_addr + size};
- std::vector<std::shared_ptr<TSurface>> surfaces;
+ std::vector<TSurface> surfaces;
for (auto& pair : boost::make_iterator_range(registered_surfaces.equal_range(interval))) {
- surfaces.push_back(*pair.second.begin());
+ for (auto& s : pair.second) {
+ if (!s || !s->IsRegistered()) {
+ continue;
+ }
+ surfaces.push_back(s);
+ }
}
return surfaces;
}
- void ReserveSurface(const SurfaceParams& params, std::shared_ptr<TSurface> surface) {
+ void RegisterInnerCache(TSurface& surface) {
+ GPUVAddr start = surface->GetGpuAddr() >> inner_cache_page_bits;
+ const GPUVAddr end = (surface->GetGpuAddrEnd() - 1) >> inner_cache_page_bits;
+ while (start <= end) {
+ inner_cache[start].push_back(surface);
+ start++;
+ }
+ }
+
+ void UnregisterInnerCache(TSurface& surface) {
+ GPUVAddr start = surface->GetGpuAddr() >> inner_cache_page_bits;
+ const GPUVAddr end = (surface->GetGpuAddrEnd() - 1) >> inner_cache_page_bits;
+ while (start <= end) {
+ inner_cache[start].remove(surface);
+ start++;
+ }
+ }
+
+ std::vector<TSurface> GetSurfacesInRegionInner(const GPUVAddr gpu_addr, const std::size_t size) {
+ if (size == 0) {
+ return {};
+ }
+ const GPUVAddr gpu_addr_end = gpu_addr + size;
+ GPUVAddr start = gpu_addr >> inner_cache_page_bits;
+ const GPUVAddr end = (gpu_addr_end - 1) >> inner_cache_page_bits;
+ std::vector<TSurface> surfaces;
+ while (start <= end) {
+ std::list<TSurface>& list = inner_cache[start];
+ for (auto& s : list) {
+ if (!s->IsPicked() && s->Overlaps(gpu_addr, gpu_addr_end)) {
+ s->MarkAsPicked(true);
+ surfaces.push_back(s);
+ }
+ }
+ start++;
+ }
+ for (auto& s : surfaces) {
+ s->MarkAsPicked(false);
+ }
+ return surfaces;
+ }
+
+ void ReserveSurface(const SurfaceParams& params, TSurface surface) {
surface_reserve[params].push_back(std::move(surface));
}
- std::shared_ptr<TSurface> TryGetReservedSurface(const SurfaceParams& params) {
+ TSurface TryGetReservedSurface(const SurfaceParams& params) {
auto search{surface_reserve.find(params)};
if (search == surface_reserve.end()) {
return {};
@@ -247,21 +505,41 @@ private:
return {};
}
- IntervalType GetSurfaceInterval(std::shared_ptr<TSurface> surface) const {
- return IntervalType::right_open(surface->GetCacheAddr(),
- surface->GetCacheAddr() + surface->GetSizeInBytes());
+ IntervalType GetInterval(const void* host_ptr, const std::size_t size) const {
+ const CacheAddr addr = ToCacheAddr(host_ptr);
+ return IntervalType::right_open(addr, addr + size);
}
+ struct RenderInfo {
+ RenderTargetConfig config;
+ TSurface target;
+ TView view;
+ };
+
+ struct DepthBufferInfo {
+ TSurface target;
+ TView view;
+ };
+
VideoCore::RasterizerInterface& rasterizer;
+ Tegra::MemoryManager* memory_manager;
u64 ticks{};
IntervalMap registered_surfaces;
+ static constexpr u64 inner_cache_page_bits{20};
+ static constexpr u64 inner_cache_page_size{1 << inner_cache_page_bits};
+ std::unordered_map<GPUVAddr, std::list<TSurface>> inner_cache;
+
/// The surface reserve is a "backup" cache, this is where we put unique surfaces that have
/// previously been used. This is to prevent surfaces from being constantly created and
/// destroyed when used with different surface parameters.
- std::unordered_map<SurfaceParams, std::list<std::shared_ptr<TSurface>>> surface_reserve;
+ std::unordered_map<SurfaceParams, std::list<TSurface>> surface_reserve;
+ std::array<RenderInfo, Tegra::Engines::Maxwell3D::Regs::NumRenderTargets> render_targets;
+ DepthBufferInfo depth_buffer;
+
+ std::vector<u8> staging_buffer;
};
} // namespace VideoCommon