From 620a1c6956207a7156ac47c0c949e64b73186db5 Mon Sep 17 00:00:00 2001 From: minjaesong Date: Wed, 16 Aug 2023 21:56:04 +0900 Subject: [PATCH] GL_UNSIGNED_SHORT works just fine on macOS?! --- src/com/badlogic/gdx/graphics/glutils/Float16FrameBuffer.java | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/com/badlogic/gdx/graphics/glutils/Float16FrameBuffer.java b/src/com/badlogic/gdx/graphics/glutils/Float16FrameBuffer.java index 4057f13b4..4e4560c30 100644 --- a/src/com/badlogic/gdx/graphics/glutils/Float16FrameBuffer.java +++ b/src/com/badlogic/gdx/graphics/glutils/Float16FrameBuffer.java @@ -35,7 +35,7 @@ public class Float16FrameBuffer extends FrameBuffer { public Float16FrameBuffer (int width, int height, boolean hasDepth) { if (App.isAppleM) { // disable float framebuffer for Apple M chips FrameBufferBuilder bufferBuilder = new FrameBufferBuilder(width, height); - bufferBuilder.addColorTextureAttachment(GL20.GL_RGBA, GL20.GL_RGBA, GL20.GL_UNSIGNED_BYTE); + bufferBuilder.addColorTextureAttachment(GL20.GL_RGBA, GL20.GL_RGBA, GL20.GL_UNSIGNED_SHORT); // but 16bpp int works perfectly?! if (hasDepth) bufferBuilder.addBasicDepthRenderBuffer(); this.bufferBuilder = bufferBuilder; }