diff --git a/desmume/src/OGLRender.cpp b/desmume/src/OGLRender.cpp index 767d49114..6df113e5f 100644 --- a/desmume/src/OGLRender.cpp +++ b/desmume/src/OGLRender.cpp @@ -50,7 +50,7 @@ static void ENDGL() { #include #include #else -#ifdef DESMUME_COCOA +#ifdef __APPLE__ #include #include #else @@ -100,12 +100,12 @@ static u32 textureFormat=0, texturePalette=0; #ifdef _WIN32 #define INITOGLEXT(x,y) y = (x)wglGetProcAddress(#y); -#elif !defined(DESMUME_COCOA) +#elif !defined(__APPLE__) #include #define INITOGLEXT(x,y) y = (x)glXGetProcAddress((const GLubyte *) #y); #endif -#ifndef DESMUME_COCOA +#ifndef __APPLE__ OGLEXT(PFNGLCREATESHADERPROC,glCreateShader) //zero: i dont understand this at all. my glext.h has the wrong thing declared here... so I have to do it myself typedef void (APIENTRYP X_PFNGLGETSHADERSOURCEPROC) (GLuint shader, GLsizei bufSize, const GLchar **source, GLsizei *length); @@ -415,7 +415,7 @@ static char OGLInit(void) glBlendFunc(GL_SRC_ALPHA, GL_ONE_MINUS_SRC_ALPHA); -#ifndef DESMUME_COCOA +#ifndef __APPLE__ INITOGLEXT(PFNGLCREATESHADERPROC,glCreateShader) INITOGLEXT(X_PFNGLGETSHADERSOURCEPROC,glShaderSource) INITOGLEXT(PFNGLCOMPILESHADERPROC,glCompileShader)