| // Copyright (c) 2012 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "gpu/command_buffer/tests/gl_manager.h" |
| |
| #include <GLES2/gl2.h> |
| #include <GLES2/gl2ext.h> |
| #include <GLES2/gl2extchromium.h> |
| #include <stddef.h> |
| #include <stdint.h> |
| |
| #include <vector> |
| |
| #include "base/at_exit.h" |
| #include "base/bind.h" |
| #include "base/command_line.h" |
| #include "base/memory/ptr_util.h" |
| #include "base/memory/ref_counted_memory.h" |
| #include "build/build_config.h" |
| #include "gpu/command_buffer/client/gles2_cmd_helper.h" |
| #include "gpu/command_buffer/client/gles2_implementation.h" |
| #include "gpu/command_buffer/client/gles2_lib.h" |
| #include "gpu/command_buffer/client/shared_memory_limits.h" |
| #include "gpu/command_buffer/client/transfer_buffer.h" |
| #include "gpu/command_buffer/common/constants.h" |
| #include "gpu/command_buffer/common/gles2_cmd_utils.h" |
| #include "gpu/command_buffer/common/sync_token.h" |
| #include "gpu/command_buffer/service/command_buffer_service.h" |
| #include "gpu/command_buffer/service/command_executor.h" |
| #include "gpu/command_buffer/service/context_group.h" |
| #include "gpu/command_buffer/service/gl_context_virtual.h" |
| #include "gpu/command_buffer/service/gles2_cmd_decoder.h" |
| #include "gpu/command_buffer/service/image_manager.h" |
| #include "gpu/command_buffer/service/mailbox_manager_impl.h" |
| #include "gpu/command_buffer/service/memory_tracking.h" |
| #include "gpu/command_buffer/service/sync_point_manager.h" |
| #include "gpu/command_buffer/service/transfer_buffer_manager.h" |
| #include "testing/gtest/include/gtest/gtest.h" |
| #include "ui/gfx/buffer_format_util.h" |
| #include "ui/gfx/gpu_memory_buffer.h" |
| #include "ui/gl/gl_context.h" |
| #include "ui/gl/gl_image_ref_counted_memory.h" |
| #include "ui/gl/gl_share_group.h" |
| #include "ui/gl/gl_surface.h" |
| #include "ui/gl/init/gl_factory.h" |
| |
| #if defined(OS_MACOSX) |
| #include "ui/gfx/mac/io_surface.h" |
| #include "ui/gl/gl_image_io_surface.h" |
| #endif |
| |
| namespace gpu { |
| namespace { |
| |
| uint64_t g_next_command_buffer_id = 0; |
| |
| class GpuMemoryBufferImpl : public gfx::GpuMemoryBuffer { |
| public: |
| GpuMemoryBufferImpl(base::RefCountedBytes* bytes, |
| const gfx::Size& size, |
| gfx::BufferFormat format) |
| : mapped_(false), bytes_(bytes), size_(size), format_(format) {} |
| |
| static GpuMemoryBufferImpl* FromClientBuffer(ClientBuffer buffer) { |
| return reinterpret_cast<GpuMemoryBufferImpl*>(buffer); |
| } |
| |
| // Overridden from gfx::GpuMemoryBuffer: |
| bool Map() override { |
| DCHECK(!mapped_); |
| mapped_ = true; |
| return true; |
| } |
| void* memory(size_t plane) override { |
| DCHECK(mapped_); |
| DCHECK_LT(plane, gfx::NumberOfPlanesForBufferFormat(format_)); |
| return reinterpret_cast<uint8_t*>(&bytes_->data().front()) + |
| gfx::BufferOffsetForBufferFormat(size_, format_, plane); |
| } |
| void Unmap() override { |
| DCHECK(mapped_); |
| mapped_ = false; |
| } |
| gfx::Size GetSize() const override { return size_; } |
| gfx::BufferFormat GetFormat() const override { return format_; } |
| int stride(size_t plane) const override { |
| DCHECK_LT(plane, gfx::NumberOfPlanesForBufferFormat(format_)); |
| return gfx::RowSizeForBufferFormat(size_.width(), format_, plane); |
| } |
| gfx::GpuMemoryBufferId GetId() const override { |
| NOTREACHED(); |
| return gfx::GpuMemoryBufferId(0); |
| } |
| gfx::GpuMemoryBufferHandle GetHandle() const override { |
| NOTREACHED(); |
| return gfx::GpuMemoryBufferHandle(); |
| } |
| ClientBuffer AsClientBuffer() override { |
| return reinterpret_cast<ClientBuffer>(this); |
| } |
| |
| base::RefCountedBytes* bytes() { return bytes_.get(); } |
| |
| private: |
| bool mapped_; |
| scoped_refptr<base::RefCountedBytes> bytes_; |
| const gfx::Size size_; |
| gfx::BufferFormat format_; |
| }; |
| |
| #if defined(OS_MACOSX) |
| class IOSurfaceGpuMemoryBuffer : public gfx::GpuMemoryBuffer { |
| public: |
| IOSurfaceGpuMemoryBuffer(const gfx::Size& size, gfx::BufferFormat format) |
| : mapped_(false), size_(size), format_(format) { |
| iosurface_ = gfx::CreateIOSurface(size, gfx::BufferFormat::BGRA_8888); |
| } |
| |
| ~IOSurfaceGpuMemoryBuffer() override { |
| CFRelease(iosurface_); |
| } |
| |
| static IOSurfaceGpuMemoryBuffer* FromClientBuffer(ClientBuffer buffer) { |
| return reinterpret_cast<IOSurfaceGpuMemoryBuffer*>(buffer); |
| } |
| |
| // Overridden from gfx::GpuMemoryBuffer: |
| bool Map() override { |
| DCHECK(!mapped_); |
| mapped_ = true; |
| return true; |
| } |
| void* memory(size_t plane) override { |
| DCHECK(mapped_); |
| DCHECK_LT(plane, gfx::NumberOfPlanesForBufferFormat(format_)); |
| return IOSurfaceGetBaseAddressOfPlane(iosurface_, plane); |
| } |
| void Unmap() override { |
| DCHECK(mapped_); |
| mapped_ = false; |
| } |
| gfx::Size GetSize() const override { return size_; } |
| gfx::BufferFormat GetFormat() const override { return format_; } |
| int stride(size_t plane) const override { |
| DCHECK_LT(plane, gfx::NumberOfPlanesForBufferFormat(format_)); |
| return IOSurfaceGetWidthOfPlane(iosurface_, plane); |
| } |
| gfx::GpuMemoryBufferId GetId() const override { |
| NOTREACHED(); |
| return gfx::GpuMemoryBufferId(0); |
| } |
| gfx::GpuMemoryBufferHandle GetHandle() const override { |
| NOTREACHED(); |
| return gfx::GpuMemoryBufferHandle(); |
| } |
| ClientBuffer AsClientBuffer() override { |
| return reinterpret_cast<ClientBuffer>(this); |
| } |
| |
| IOSurfaceRef iosurface() { return iosurface_; } |
| |
| private: |
| bool mapped_; |
| IOSurfaceRef iosurface_; |
| const gfx::Size size_; |
| gfx::BufferFormat format_; |
| }; |
| #endif // defined(OS_MACOSX) |
| |
| } // namespace |
| |
| int GLManager::use_count_; |
| scoped_refptr<gfx::GLShareGroup>* GLManager::base_share_group_; |
| scoped_refptr<gfx::GLSurface>* GLManager::base_surface_; |
| scoped_refptr<gfx::GLContext>* GLManager::base_context_; |
| |
| GLManager::Options::Options() |
| : size(4, 4), |
| sync_point_manager(NULL), |
| share_group_manager(NULL), |
| share_mailbox_manager(NULL), |
| virtual_manager(NULL), |
| bind_generates_resource(false), |
| lose_context_when_out_of_memory(false), |
| context_lost_allowed(false), |
| context_type(gles2::CONTEXT_TYPE_OPENGLES2), |
| force_shader_name_hashing(false) {} |
| |
| GLManager::GLManager() |
| : sync_point_manager_(nullptr), |
| context_lost_allowed_(false), |
| pause_commands_(false), |
| paused_order_num_(0), |
| command_buffer_id_( |
| CommandBufferId::FromUnsafeValue(g_next_command_buffer_id++)), |
| next_fence_sync_release_(1) { |
| SetupBaseContext(); |
| } |
| |
| GLManager::~GLManager() { |
| --use_count_; |
| if (!use_count_) { |
| if (base_share_group_) { |
| delete base_context_; |
| base_context_ = NULL; |
| } |
| if (base_surface_) { |
| delete base_surface_; |
| base_surface_ = NULL; |
| } |
| if (base_context_) { |
| delete base_context_; |
| base_context_ = NULL; |
| } |
| } |
| } |
| |
| std::unique_ptr<gfx::GpuMemoryBuffer> GLManager::CreateGpuMemoryBuffer( |
| const gfx::Size& size, |
| gfx::BufferFormat format) { |
| #if defined(OS_MACOSX) |
| if (use_iosurface_memory_buffers_) { |
| return base::WrapUnique<gfx::GpuMemoryBuffer>( |
| new IOSurfaceGpuMemoryBuffer(size, format)); |
| } |
| #endif // defined(OS_MACOSX) |
| std::vector<uint8_t> data(gfx::BufferSizeForBufferFormat(size, format), 0); |
| scoped_refptr<base::RefCountedBytes> bytes(new base::RefCountedBytes(data)); |
| return base::WrapUnique<gfx::GpuMemoryBuffer>( |
| new GpuMemoryBufferImpl(bytes.get(), size, format)); |
| } |
| |
| void GLManager::Initialize(const GLManager::Options& options) { |
| InitializeWithCommandLine(options, *base::CommandLine::ForCurrentProcess()); |
| } |
| |
| void GLManager::InitializeWithCommandLine( |
| const GLManager::Options& options, |
| const base::CommandLine& command_line) { |
| const int32_t kCommandBufferSize = 1024 * 1024; |
| const size_t kStartTransferBufferSize = 4 * 1024 * 1024; |
| const size_t kMinTransferBufferSize = 1 * 256 * 1024; |
| const size_t kMaxTransferBufferSize = 16 * 1024 * 1024; |
| |
| context_lost_allowed_ = options.context_lost_allowed; |
| |
| gles2::MailboxManager* mailbox_manager = NULL; |
| if (options.share_mailbox_manager) { |
| mailbox_manager = options.share_mailbox_manager->mailbox_manager(); |
| } else if (options.share_group_manager) { |
| mailbox_manager = options.share_group_manager->mailbox_manager(); |
| } |
| |
| gfx::GLShareGroup* share_group = NULL; |
| if (options.share_group_manager) { |
| share_group = options.share_group_manager->share_group(); |
| } else if (options.share_mailbox_manager) { |
| share_group = options.share_mailbox_manager->share_group(); |
| } |
| |
| gles2::ContextGroup* context_group = NULL; |
| scoped_refptr<gles2::ShareGroup> client_share_group; |
| if (options.share_group_manager) { |
| context_group = options.share_group_manager->decoder_->GetContextGroup(); |
| client_share_group = |
| options.share_group_manager->gles2_implementation()->share_group(); |
| } |
| |
| gfx::GLContext* real_gl_context = NULL; |
| if (options.virtual_manager) { |
| real_gl_context = options.virtual_manager->context(); |
| } |
| |
| mailbox_manager_ = |
| mailbox_manager ? mailbox_manager : new gles2::MailboxManagerImpl; |
| share_group_ = |
| share_group ? share_group : new gfx::GLShareGroup; |
| |
| gfx::GpuPreference gpu_preference(gfx::PreferDiscreteGpu); |
| std::vector<int32_t> attribs; |
| gles2::ContextCreationAttribHelper attrib_helper; |
| attrib_helper.red_size = 8; |
| attrib_helper.green_size = 8; |
| attrib_helper.blue_size = 8; |
| attrib_helper.alpha_size = 8; |
| attrib_helper.depth_size = 16; |
| attrib_helper.stencil_size = 8; |
| attrib_helper.context_type = options.context_type; |
| |
| attrib_helper.Serialize(&attribs); |
| |
| if (!context_group) { |
| GpuDriverBugWorkarounds gpu_driver_bug_workaround(&command_line); |
| scoped_refptr<gles2::FeatureInfo> feature_info = |
| new gles2::FeatureInfo(command_line, gpu_driver_bug_workaround); |
| context_group = new gles2::ContextGroup( |
| gpu_preferences_, mailbox_manager_.get(), NULL, |
| new gpu::gles2::ShaderTranslatorCache(gpu_preferences_), |
| new gpu::gles2::FramebufferCompletenessCache, feature_info, |
| options.bind_generates_resource); |
| } |
| |
| decoder_.reset(::gpu::gles2::GLES2Decoder::Create(context_group)); |
| if (options.force_shader_name_hashing) { |
| decoder_->SetForceShaderNameHashingForTest(true); |
| } |
| command_buffer_.reset(new CommandBufferService( |
| decoder_->GetContextGroup()->transfer_buffer_manager())); |
| |
| executor_.reset(new CommandExecutor(command_buffer_.get(), decoder_.get(), |
| decoder_.get())); |
| |
| decoder_->set_engine(executor_.get()); |
| |
| surface_ = gl::init::CreateOffscreenGLSurface(gfx::Size()); |
| ASSERT_TRUE(surface_.get() != NULL) << "could not create offscreen surface"; |
| |
| if (base_context_) { |
| context_ = scoped_refptr<gfx::GLContext>(new gpu::GLContextVirtual( |
| share_group_.get(), base_context_->get(), decoder_->AsWeakPtr())); |
| ASSERT_TRUE(context_->Initialize( |
| surface_.get(), gfx::PreferIntegratedGpu)); |
| } else { |
| if (real_gl_context) { |
| context_ = scoped_refptr<gfx::GLContext>(new gpu::GLContextVirtual( |
| share_group_.get(), real_gl_context, decoder_->AsWeakPtr())); |
| ASSERT_TRUE(context_->Initialize( |
| surface_.get(), gfx::PreferIntegratedGpu)); |
| } else { |
| context_ = gl::init::CreateGLContext(share_group_.get(), surface_.get(), |
| gpu_preference); |
| } |
| } |
| ASSERT_TRUE(context_.get() != NULL) << "could not create GL context"; |
| |
| ASSERT_TRUE(context_->MakeCurrent(surface_.get())); |
| |
| if (!decoder_->Initialize(surface_.get(), context_.get(), true, options.size, |
| ::gpu::gles2::DisallowedFeatures(), attribs)) { |
| return; |
| } |
| |
| if (options.sync_point_manager) { |
| sync_point_manager_ = options.sync_point_manager; |
| sync_point_order_data_ = SyncPointOrderData::Create(); |
| sync_point_client_ = sync_point_manager_->CreateSyncPointClient( |
| sync_point_order_data_, GetNamespaceID(), GetCommandBufferID()); |
| |
| decoder_->SetFenceSyncReleaseCallback( |
| base::Bind(&GLManager::OnFenceSyncRelease, base::Unretained(this))); |
| decoder_->SetWaitFenceSyncCallback( |
| base::Bind(&GLManager::OnWaitFenceSync, base::Unretained(this))); |
| } else { |
| sync_point_manager_ = nullptr; |
| sync_point_order_data_ = nullptr; |
| sync_point_client_ = nullptr; |
| } |
| |
| command_buffer_->SetPutOffsetChangeCallback( |
| base::Bind(&GLManager::PumpCommands, base::Unretained(this))); |
| command_buffer_->SetGetBufferChangeCallback( |
| base::Bind(&GLManager::GetBufferChanged, base::Unretained(this))); |
| |
| // Create the GLES2 helper, which writes the command buffer protocol. |
| gles2_helper_.reset(new gles2::GLES2CmdHelper(command_buffer_.get())); |
| ASSERT_TRUE(gles2_helper_->Initialize(kCommandBufferSize)); |
| |
| // Create a transfer buffer. |
| transfer_buffer_.reset(new TransferBuffer(gles2_helper_.get())); |
| |
| // Create the object exposing the OpenGL API. |
| const bool support_client_side_arrays = true; |
| gles2_implementation_.reset(new gles2::GLES2Implementation( |
| gles2_helper_.get(), std::move(client_share_group), |
| transfer_buffer_.get(), options.bind_generates_resource, |
| options.lose_context_when_out_of_memory, support_client_side_arrays, |
| this)); |
| |
| ASSERT_TRUE(gles2_implementation_->Initialize( |
| kStartTransferBufferSize, kMinTransferBufferSize, kMaxTransferBufferSize, |
| SharedMemoryLimits::kNoLimit)) |
| << "Could not init GLES2Implementation"; |
| |
| MakeCurrent(); |
| } |
| |
| void GLManager::SetupBaseContext() { |
| if (use_count_) { |
| #if defined(OS_ANDROID) |
| base_share_group_ = new scoped_refptr<gfx::GLShareGroup>( |
| new gfx::GLShareGroup); |
| gfx::Size size(4, 4); |
| base_surface_ = new scoped_refptr<gfx::GLSurface>( |
| gl::init::CreateOffscreenGLSurface(size)); |
| gfx::GpuPreference gpu_preference(gfx::PreferDiscreteGpu); |
| base_context_ = |
| new scoped_refptr<gfx::GLContext>(gl::init::CreateGLContext( |
| base_share_group_->get(), base_surface_->get(), gpu_preference)); |
| #endif |
| } |
| ++use_count_; |
| } |
| |
| void GLManager::OnFenceSyncRelease(uint64_t release) { |
| DCHECK(sync_point_client_); |
| DCHECK(!sync_point_client_->client_state()->IsFenceSyncReleased(release)); |
| sync_point_client_->ReleaseFenceSync(release); |
| } |
| |
| bool GLManager::OnWaitFenceSync(gpu::CommandBufferNamespace namespace_id, |
| gpu::CommandBufferId command_buffer_id, |
| uint64_t release) { |
| DCHECK(sync_point_client_); |
| scoped_refptr<gpu::SyncPointClientState> release_state = |
| sync_point_manager_->GetSyncPointClientState(namespace_id, |
| command_buffer_id); |
| if (!release_state) |
| return true; |
| |
| // GLManager does not support being multithreaded at this point, so the fence |
| // sync must be released by the time wait is called. |
| DCHECK(release_state->IsFenceSyncReleased(release)); |
| return true; |
| } |
| |
| void GLManager::MakeCurrent() { |
| ::gles2::SetGLContext(gles2_implementation_.get()); |
| } |
| |
| void GLManager::SetSurface(gfx::GLSurface* surface) { |
| decoder_->SetSurface(surface); |
| } |
| |
| void GLManager::Destroy() { |
| if (gles2_implementation_.get()) { |
| MakeCurrent(); |
| EXPECT_TRUE(glGetError() == GL_NONE); |
| gles2_implementation_->Flush(); |
| gles2_implementation_.reset(); |
| } |
| transfer_buffer_.reset(); |
| gles2_helper_.reset(); |
| command_buffer_.reset(); |
| sync_point_manager_ = nullptr; |
| sync_point_client_ = nullptr; |
| if (sync_point_order_data_) { |
| sync_point_order_data_->Destroy(); |
| sync_point_order_data_ = nullptr; |
| } |
| if (decoder_.get()) { |
| bool have_context = decoder_->GetGLContext() && |
| decoder_->GetGLContext()->MakeCurrent(surface_.get()); |
| decoder_->Destroy(have_context); |
| decoder_.reset(); |
| } |
| } |
| |
| const GpuDriverBugWorkarounds& GLManager::workarounds() const { |
| return decoder_->GetContextGroup()->feature_info()->workarounds(); |
| } |
| |
| void GLManager::PumpCommands() { |
| if (!decoder_->MakeCurrent()) { |
| command_buffer_->SetContextLostReason(decoder_->GetContextLostReason()); |
| command_buffer_->SetParseError(::gpu::error::kLostContext); |
| return; |
| } |
| uint32_t order_num = 0; |
| if (sync_point_manager_) { |
| // If sync point manager is supported, assign order numbers to commands. |
| if (paused_order_num_) { |
| // Was previous paused, continue to process the order number. |
| order_num = paused_order_num_; |
| paused_order_num_ = 0; |
| } else { |
| order_num = sync_point_order_data_->GenerateUnprocessedOrderNumber( |
| sync_point_manager_); |
| } |
| sync_point_order_data_->BeginProcessingOrderNumber(order_num); |
| } |
| |
| if (pause_commands_) { |
| // Do not process commands, simply store the current order number. |
| paused_order_num_ = order_num; |
| |
| sync_point_order_data_->PauseProcessingOrderNumber(order_num); |
| return; |
| } |
| |
| executor_->PutChanged(); |
| ::gpu::CommandBuffer::State state = command_buffer_->GetLastState(); |
| if (!context_lost_allowed_) { |
| ASSERT_EQ(::gpu::error::kNoError, state.error); |
| } |
| |
| if (sync_point_manager_) { |
| // Finish processing order number here. |
| sync_point_order_data_->FinishProcessingOrderNumber(order_num); |
| } |
| } |
| |
| bool GLManager::GetBufferChanged(int32_t transfer_buffer_id) { |
| return executor_->SetGetBuffer(transfer_buffer_id); |
| } |
| |
| void GLManager::SetGpuControlClient(GpuControlClient*) { |
| // The client is not currently called, so don't store it. |
| } |
| |
| Capabilities GLManager::GetCapabilities() { |
| return decoder_->GetCapabilities(); |
| } |
| |
| int32_t GLManager::CreateImage(ClientBuffer buffer, |
| size_t width, |
| size_t height, |
| unsigned internalformat) { |
| gfx::Size size(width, height); |
| scoped_refptr<gl::GLImage> gl_image; |
| |
| #if defined(OS_MACOSX) |
| if (use_iosurface_memory_buffers_) { |
| IOSurfaceGpuMemoryBuffer* gpu_memory_buffer = |
| IOSurfaceGpuMemoryBuffer::FromClientBuffer(buffer); |
| scoped_refptr<gl::GLImageIOSurface> image( |
| new gl::GLImageIOSurface(size, internalformat)); |
| if (!image->Initialize(gpu_memory_buffer->iosurface(), |
| gfx::GenericSharedMemoryId(1), |
| gfx::BufferFormat::BGRA_8888)) { |
| return -1; |
| } |
| gl_image = image; |
| } |
| #endif // defined(OS_MACOSX) |
| if (!gl_image) { |
| GpuMemoryBufferImpl* gpu_memory_buffer = |
| GpuMemoryBufferImpl::FromClientBuffer(buffer); |
| |
| scoped_refptr<gl::GLImageRefCountedMemory> image( |
| new gl::GLImageRefCountedMemory(size, internalformat)); |
| if (!image->Initialize(gpu_memory_buffer->bytes(), |
| gpu_memory_buffer->GetFormat())) { |
| return -1; |
| } |
| gl_image = image; |
| } |
| |
| static int32_t next_id = 1; |
| int32_t new_id = next_id++; |
| gpu::gles2::ImageManager* image_manager = decoder_->GetImageManager(); |
| DCHECK(image_manager); |
| image_manager->AddImage(gl_image.get(), new_id); |
| return new_id; |
| } |
| |
| int32_t GLManager::CreateGpuMemoryBufferImage(size_t width, |
| size_t height, |
| unsigned internalformat, |
| unsigned usage) { |
| DCHECK_EQ(usage, static_cast<unsigned>(GL_READ_WRITE_CHROMIUM)); |
| std::unique_ptr<gfx::GpuMemoryBuffer> buffer = CreateGpuMemoryBuffer( |
| gfx::Size(width, height), gfx::BufferFormat::RGBA_8888); |
| return CreateImage(buffer->AsClientBuffer(), width, height, internalformat); |
| } |
| |
| void GLManager::DestroyImage(int32_t id) { |
| gpu::gles2::ImageManager* image_manager = decoder_->GetImageManager(); |
| DCHECK(image_manager); |
| image_manager->RemoveImage(id); |
| } |
| |
| void GLManager::SignalQuery(uint32_t query, const base::Closure& callback) { |
| NOTIMPLEMENTED(); |
| } |
| |
| void GLManager::SetLock(base::Lock*) { |
| NOTIMPLEMENTED(); |
| } |
| |
| void GLManager::EnsureWorkVisible() { |
| // This is only relevant for out-of-process command buffers. |
| } |
| |
| gpu::CommandBufferNamespace GLManager::GetNamespaceID() const { |
| return gpu::CommandBufferNamespace::IN_PROCESS; |
| } |
| |
| CommandBufferId GLManager::GetCommandBufferID() const { |
| return command_buffer_id_; |
| } |
| |
| int32_t GLManager::GetExtraCommandBufferData() const { |
| return 0; |
| } |
| |
| uint64_t GLManager::GenerateFenceSyncRelease() { |
| return next_fence_sync_release_++; |
| } |
| |
| bool GLManager::IsFenceSyncRelease(uint64_t release) { |
| return release > 0 && release < next_fence_sync_release_; |
| } |
| |
| bool GLManager::IsFenceSyncFlushed(uint64_t release) { |
| return IsFenceSyncRelease(release); |
| } |
| |
| bool GLManager::IsFenceSyncFlushReceived(uint64_t release) { |
| return IsFenceSyncRelease(release); |
| } |
| |
| void GLManager::SignalSyncToken(const gpu::SyncToken& sync_token, |
| const base::Closure& callback) { |
| if (sync_point_manager_) { |
| scoped_refptr<gpu::SyncPointClientState> release_state = |
| sync_point_manager_->GetSyncPointClientState( |
| sync_token.namespace_id(), sync_token.command_buffer_id()); |
| |
| if (release_state) { |
| sync_point_client_->WaitOutOfOrder(release_state.get(), |
| sync_token.release_count(), callback); |
| return; |
| } |
| } |
| |
| // Something went wrong, just run the callback now. |
| callback.Run(); |
| } |
| |
| bool GLManager::CanWaitUnverifiedSyncToken(const gpu::SyncToken* sync_token) { |
| return false; |
| } |
| |
| } // namespace gpu |