protot/3rdparty/bgfx/examples/common/nanovg/nanovg_bgfx.cpp

1205 lines
32 KiB
C++
Raw Permalink Normal View History

2016-08-29 22:31:11 +02:00
/*
2017-04-11 08:16:10 +02:00
* Copyright 2011-2017 Branimir Karadzic. All rights reserved.
2016-08-29 22:31:11 +02:00
* License: https://github.com/bkaradzic/bgfx#license-bsd-2-clause
*/
//
// Copyright (c) 2009-2013 Mikko Mononen memon@inside.org
//
// This software is provided 'as-is', without any express or implied
// warranty. In no event will the authors be held liable for any damages
// arising from the use of this software.
// Permission is granted to anyone to use this software for any purpose,
// including commercial applications, and to alter it and redistribute it
// freely, subject to the following restrictions:
// 1. The origin of this software must not be misrepresented; you must not
// claim that you wrote the original software. If you use this software
// in a product, an acknowledgment in the product documentation would be
// appreciated but is not required.
// 2. Altered source versions must be plainly marked as such, and must not be
// misrepresented as being the original software.
// 3. This notice may not be removed or altered from any source distribution.
//
#define NVG_ANTIALIAS 1
#include <stdlib.h>
#include <math.h>
#include "nanovg.h"
#include <bgfx/bgfx.h>
2017-04-11 08:16:10 +02:00
#include <bgfx/embedded_shader.h>
2016-08-29 22:31:11 +02:00
#include <bx/bx.h>
#include <bx/allocator.h>
#include <bx/crtimpl.h>
2016-11-10 21:53:08 +01:00
#include <bx/uint32_t.h>
2016-08-29 22:31:11 +02:00
BX_PRAGMA_DIAGNOSTIC_IGNORED_MSVC(4244); // warning C4244: '=' : conversion from '' to '', possible loss of data
#include "vs_nanovg_fill.bin.h"
#include "fs_nanovg_fill.bin.h"
2017-04-11 08:16:10 +02:00
static const bgfx::EmbeddedShader s_embeddedShaders[] =
{
BGFX_EMBEDDED_SHADER(vs_nanovg_fill),
BGFX_EMBEDDED_SHADER(fs_nanovg_fill),
BGFX_EMBEDDED_SHADER_END()
};
namespace
{
2016-08-29 22:31:11 +02:00
static bgfx::VertexDecl s_nvgDecl;
enum GLNVGshaderType
{
NSVG_SHADER_FILLGRAD,
NSVG_SHADER_FILLIMG,
NSVG_SHADER_SIMPLE,
NSVG_SHADER_IMG
};
// These are additional flags on top of NVGimageFlags.
enum NVGimageFlagsGL {
NVG_IMAGE_NODELETE = 1<<16, // Do not delete GL texture handle.
};
struct GLNVGtexture
{
bgfx::TextureHandle id;
int width, height;
int type;
int flags;
};
2017-04-11 08:16:10 +02:00
struct GLNVGblend
{
uint64_t srcRGB;
uint64_t dstRGB;
uint64_t srcAlpha;
uint64_t dstAlpha;
};
2016-08-29 22:31:11 +02:00
enum GLNVGcallType
{
GLNVG_FILL,
GLNVG_CONVEXFILL,
GLNVG_STROKE,
GLNVG_TRIANGLES,
};
struct GLNVGcall
{
int type;
int image;
int pathOffset;
int pathCount;
int vertexOffset;
int vertexCount;
int uniformOffset;
2017-04-11 08:16:10 +02:00
GLNVGblend blendFunc;
2016-08-29 22:31:11 +02:00
};
struct GLNVGpath
{
int fillOffset;
int fillCount;
int strokeOffset;
int strokeCount;
};
struct GLNVGfragUniforms
{
float scissorMat[12]; // matrices are actually 3 vec4s
float paintMat[12];
NVGcolor innerCol;
NVGcolor outerCol;
// u_scissorExtScale
float scissorExt[2];
float scissorScale[2];
// u_extentRadius
float extent[2];
float radius;
// u_params
float feather;
float strokeMult;
float texType;
float type;
};
struct GLNVGcontext
{
bx::AllocatorI* m_allocator;
bgfx::ProgramHandle prog;
bgfx::UniformHandle u_scissorMat;
bgfx::UniformHandle u_paintMat;
bgfx::UniformHandle u_innerCol;
bgfx::UniformHandle u_outerCol;
bgfx::UniformHandle u_viewSize;
bgfx::UniformHandle u_scissorExtScale;
bgfx::UniformHandle u_extentRadius;
bgfx::UniformHandle u_params;
bgfx::UniformHandle u_halfTexel;
bgfx::UniformHandle s_tex;
uint64_t state;
bgfx::TextureHandle th;
bgfx::TextureHandle texMissing;
bgfx::TransientVertexBuffer tvb;
uint8_t m_viewId;
struct GLNVGtexture* textures;
float view[2];
int ntextures;
int ctextures;
int textureId;
int vertBuf;
int fragSize;
int edgeAntiAlias;
// Per frame buffers
struct GLNVGcall* calls;
int ccalls;
int ncalls;
struct GLNVGpath* paths;
int cpaths;
int npaths;
struct NVGvertex* verts;
int cverts;
int nverts;
unsigned char* uniforms;
int cuniforms;
int nuniforms;
};
static struct GLNVGtexture* glnvg__allocTexture(struct GLNVGcontext* gl)
{
struct GLNVGtexture* tex = NULL;
int i;
for (i = 0; i < gl->ntextures; i++)
{
if (gl->textures[i].id.idx == bgfx::invalidHandle)
{
tex = &gl->textures[i];
break;
}
}
if (tex == NULL)
{
if (gl->ntextures+1 > gl->ctextures)
{
int old = gl->ctextures;
gl->ctextures = (gl->ctextures == 0) ? 2 : gl->ctextures*2;
gl->textures = (struct GLNVGtexture*)BX_REALLOC(gl->m_allocator, gl->textures, sizeof(struct GLNVGtexture)*gl->ctextures);
2017-04-11 08:16:10 +02:00
bx::memSet(&gl->textures[old], 0xff, (gl->ctextures-old)*sizeof(struct GLNVGtexture) );
2016-08-29 22:31:11 +02:00
if (gl->textures == NULL)
{
return NULL;
}
}
tex = &gl->textures[gl->ntextures++];
}
2017-04-11 08:16:10 +02:00
bx::memSet(tex, 0, sizeof(*tex) );
2016-08-29 22:31:11 +02:00
return tex;
}
static struct GLNVGtexture* glnvg__findTexture(struct GLNVGcontext* gl, int id)
{
int i;
for (i = 0; i < gl->ntextures; i++)
{
if (gl->textures[i].id.idx == id)
{
return &gl->textures[i];
}
}
return NULL;
}
static int glnvg__deleteTexture(struct GLNVGcontext* gl, int id)
{
for (int ii = 0; ii < gl->ntextures; ii++)
{
if (gl->textures[ii].id.idx == id)
{
if (bgfx::isValid(gl->textures[ii].id)
&& (gl->textures[ii].flags & NVG_IMAGE_NODELETE) == 0)
{
bgfx::destroyTexture(gl->textures[ii].id);
}
2017-04-11 08:16:10 +02:00
bx::memSet(&gl->textures[ii], 0, sizeof(gl->textures[ii]) );
2016-08-29 22:31:11 +02:00
gl->textures[ii].id.idx = bgfx::invalidHandle;
return 1;
}
}
return 0;
}
static int nvgRenderCreate(void* _userPtr)
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
2017-04-11 08:16:10 +02:00
bgfx::RendererType::Enum type = bgfx::getRendererType();
2016-08-29 22:31:11 +02:00
gl->prog = bgfx::createProgram(
2017-04-11 08:16:10 +02:00
bgfx::createEmbeddedShader(s_embeddedShaders, type, "vs_nanovg_fill")
, bgfx::createEmbeddedShader(s_embeddedShaders, type, "fs_nanovg_fill")
2016-08-29 22:31:11 +02:00
, true
);
const bgfx::Memory* mem = bgfx::alloc(4*4*4);
uint32_t* bgra8 = (uint32_t*)mem->data;
2017-04-11 08:16:10 +02:00
bx::memSet(bgra8, 0, 4*4*4);
2016-08-29 22:31:11 +02:00
gl->texMissing = bgfx::createTexture2D(4, 4, false, 1, bgfx::TextureFormat::BGRA8, 0, mem);
gl->u_scissorMat = bgfx::createUniform("u_scissorMat", bgfx::UniformType::Mat3);
gl->u_paintMat = bgfx::createUniform("u_paintMat", bgfx::UniformType::Mat3);
gl->u_innerCol = bgfx::createUniform("u_innerCol", bgfx::UniformType::Vec4);
gl->u_outerCol = bgfx::createUniform("u_outerCol", bgfx::UniformType::Vec4);
gl->u_viewSize = bgfx::createUniform("u_viewSize", bgfx::UniformType::Vec4);
gl->u_scissorExtScale = bgfx::createUniform("u_scissorExtScale", bgfx::UniformType::Vec4);
gl->u_extentRadius = bgfx::createUniform("u_extentRadius", bgfx::UniformType::Vec4);
gl->u_params = bgfx::createUniform("u_params", bgfx::UniformType::Vec4);
gl->s_tex = bgfx::createUniform("s_tex", bgfx::UniformType::Int1);
if (bgfx::getRendererType() == bgfx::RendererType::Direct3D9)
{
gl->u_halfTexel = bgfx::createUniform("u_halfTexel", bgfx::UniformType::Vec4);
}
else
{
gl->u_halfTexel.idx = bgfx::invalidHandle;
}
s_nvgDecl
.begin()
.add(bgfx::Attrib::Position, 2, bgfx::AttribType::Float)
.add(bgfx::Attrib::TexCoord0, 2, bgfx::AttribType::Float)
.end();
int align = 16;
gl->fragSize = sizeof(struct GLNVGfragUniforms) + align - sizeof(struct GLNVGfragUniforms) % align;
return 1;
}
static int nvgRenderCreateTexture(void* _userPtr, int _type, int _width, int _height, int _flags, const unsigned char* _rgba)
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
struct GLNVGtexture* tex = glnvg__allocTexture(gl);
if (tex == NULL)
{
return 0;
}
tex->width = _width;
tex->height = _height;
tex->type = _type;
tex->flags = _flags;
uint32_t bytesPerPixel = NVG_TEXTURE_RGBA == tex->type ? 4 : 1;
uint32_t pitch = tex->width * bytesPerPixel;
const bgfx::Memory* mem = NULL;
if (NULL != _rgba)
{
mem = bgfx::copy(_rgba, tex->height * pitch);
}
tex->id = bgfx::createTexture2D(
tex->width
, tex->height
, false
, 1
, NVG_TEXTURE_RGBA == _type ? bgfx::TextureFormat::RGBA8 : bgfx::TextureFormat::R8
, BGFX_TEXTURE_NONE
);
if (NULL != mem)
{
bgfx::updateTexture2D(
tex->id
, 0
, 0
, 0
, 0
, tex->width
, tex->height
, mem
);
}
return bgfx::isValid(tex->id) ? tex->id.idx : 0;
}
static int nvgRenderDeleteTexture(void* _userPtr, int image)
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
return glnvg__deleteTexture(gl, image);
}
static int nvgRenderUpdateTexture(void* _userPtr, int image, int x, int y, int w, int h, const unsigned char* data)
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
struct GLNVGtexture* tex = glnvg__findTexture(gl, image);
if (tex == NULL)
{
return 0;
}
uint32_t bytesPerPixel = NVG_TEXTURE_RGBA == tex->type ? 4 : 1;
uint32_t pitch = tex->width * bytesPerPixel;
2017-04-11 08:16:10 +02:00
const bgfx::Memory* mem = bgfx::alloc(w * h * bytesPerPixel);
bx::gather(mem->data, data + y * pitch + x * bytesPerPixel, w * bytesPerPixel, h, pitch);
2016-08-29 22:31:11 +02:00
bgfx::updateTexture2D(
tex->id
, 0
, 0
, x
, y
, w
, h
2017-04-11 08:16:10 +02:00
, mem
, UINT16_MAX
2016-08-29 22:31:11 +02:00
);
return 1;
}
static int nvgRenderGetTextureSize(void* _userPtr, int image, int* w, int* h)
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
struct GLNVGtexture* tex = glnvg__findTexture(gl, image);
if (!bgfx::isValid(tex->id) )
{
return 0;
}
*w = tex->width;
*h = tex->height;
return 1;
}
static void glnvg__xformToMat3x4(float* m3, float* t)
{
m3[0] = t[0];
m3[1] = t[1];
m3[2] = 0.0f;
m3[3] = 0.0f;
m3[4] = t[2];
m3[5] = t[3];
m3[6] = 0.0f;
m3[7] = 0.0f;
m3[8] = t[4];
m3[9] = t[5];
m3[10] = 1.0f;
m3[11] = 0.0f;
}
static NVGcolor glnvg__premulColor(NVGcolor c)
{
c.r *= c.a;
c.g *= c.a;
c.b *= c.a;
return c;
}
static int glnvg__convertPaint(struct GLNVGcontext* gl, struct GLNVGfragUniforms* frag, struct NVGpaint* paint,
struct NVGscissor* scissor, float width, float fringe)
{
struct GLNVGtexture* tex = NULL;
float invxform[6] = {};
2017-04-11 08:16:10 +02:00
bx::memSet(frag, 0, sizeof(*frag) );
2016-08-29 22:31:11 +02:00
frag->innerCol = glnvg__premulColor(paint->innerColor);
frag->outerCol = glnvg__premulColor(paint->outerColor);
if (scissor->extent[0] < -0.5f || scissor->extent[1] < -0.5f)
{
2017-04-11 08:16:10 +02:00
bx::memSet(frag->scissorMat, 0, sizeof(frag->scissorMat) );
2016-08-29 22:31:11 +02:00
frag->scissorExt[0] = 1.0f;
frag->scissorExt[1] = 1.0f;
frag->scissorScale[0] = 1.0f;
frag->scissorScale[1] = 1.0f;
}
else
{
nvgTransformInverse(invxform, scissor->xform);
glnvg__xformToMat3x4(frag->scissorMat, invxform);
frag->scissorExt[0] = scissor->extent[0];
frag->scissorExt[1] = scissor->extent[1];
frag->scissorScale[0] = sqrtf(scissor->xform[0]*scissor->xform[0] + scissor->xform[2]*scissor->xform[2]) / fringe;
frag->scissorScale[1] = sqrtf(scissor->xform[1]*scissor->xform[1] + scissor->xform[3]*scissor->xform[3]) / fringe;
}
2017-04-11 08:16:10 +02:00
bx::memCopy(frag->extent, paint->extent, sizeof(frag->extent) );
2016-08-29 22:31:11 +02:00
frag->strokeMult = (width*0.5f + fringe*0.5f) / fringe;
gl->th = gl->texMissing;
if (paint->image != 0)
{
tex = glnvg__findTexture(gl, paint->image);
if (tex == NULL)
{
return 0;
}
nvgTransformInverse(invxform, paint->xform);
frag->type = NSVG_SHADER_FILLIMG;
if (tex->type == NVG_TEXTURE_RGBA)
frag->texType = (tex->flags & NVG_IMAGE_PREMULTIPLIED) ? 0.0f : 1.0f;
else
frag->texType = 2.0f;
gl->th = tex->id;
}
else
{
frag->type = NSVG_SHADER_FILLGRAD;
frag->radius = paint->radius;
frag->feather = paint->feather;
nvgTransformInverse(invxform, paint->xform);
}
glnvg__xformToMat3x4(frag->paintMat, invxform);
return 1;
}
static void glnvg__mat3(float* dst, float* src)
{
dst[0] = src[0];
dst[1] = src[1];
dst[2] = src[2];
dst[3] = src[4];
dst[4] = src[5];
dst[5] = src[6];
dst[6] = src[8];
dst[7] = src[9];
dst[8] = src[10];
}
static struct GLNVGfragUniforms* nvg__fragUniformPtr(struct GLNVGcontext* gl, int i)
{
return (struct GLNVGfragUniforms*)&gl->uniforms[i];
}
static void nvgRenderSetUniforms(struct GLNVGcontext* gl, int uniformOffset, int image)
{
struct GLNVGfragUniforms* frag = nvg__fragUniformPtr(gl, uniformOffset);
float tmp[9]; // Maybe there's a way to get rid of this...
glnvg__mat3(tmp, frag->scissorMat);
bgfx::setUniform(gl->u_scissorMat, tmp);
glnvg__mat3(tmp, frag->paintMat);
bgfx::setUniform(gl->u_paintMat, tmp);
bgfx::setUniform(gl->u_innerCol, frag->innerCol.rgba);
bgfx::setUniform(gl->u_outerCol, frag->outerCol.rgba);
bgfx::setUniform(gl->u_scissorExtScale, &frag->scissorExt[0]);
bgfx::setUniform(gl->u_extentRadius, &frag->extent[0]);
bgfx::setUniform(gl->u_params, &frag->feather);
bgfx::TextureHandle handle = gl->texMissing;
if (image != 0)
{
struct GLNVGtexture* tex = glnvg__findTexture(gl, image);
if (tex != NULL)
{
handle = tex->id;
if (bgfx::isValid(gl->u_halfTexel) )
{
float halfTexel[4] = { 0.5f / tex->width, 0.5f / tex->height };
bgfx::setUniform(gl->u_halfTexel, halfTexel);
}
}
}
gl->th = handle;
}
static void nvgRenderViewport(void* _userPtr, int width, int height, float devicePixelRatio)
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
gl->view[0] = (float)width;
gl->view[1] = (float)height;
bgfx::setViewRect(gl->m_viewId, 0, 0, width * devicePixelRatio, height * devicePixelRatio);
}
static void fan(uint32_t _start, uint32_t _count)
{
uint32_t numTris = _count-2;
bgfx::TransientIndexBuffer tib;
bgfx::allocTransientIndexBuffer(&tib, numTris*3);
uint16_t* data = (uint16_t*)tib.data;
for (uint32_t ii = 0; ii < numTris; ++ii)
{
data[ii*3+0] = _start;
data[ii*3+1] = _start + ii + 1;
data[ii*3+2] = _start + ii + 2;
}
bgfx::setIndexBuffer(&tib);
}
static void glnvg__fill(struct GLNVGcontext* gl, struct GLNVGcall* call)
{
struct GLNVGpath* paths = &gl->paths[call->pathOffset];
int i, npaths = call->pathCount;
// set bindpoint for solid loc
nvgRenderSetUniforms(gl, call->uniformOffset, 0);
for (i = 0; i < npaths; i++)
{
if (2 < paths[i].fillCount)
{
bgfx::setState(0);
bgfx::setStencil(0
| BGFX_STENCIL_TEST_ALWAYS
| BGFX_STENCIL_FUNC_RMASK(0xff)
| BGFX_STENCIL_OP_FAIL_S_KEEP
| BGFX_STENCIL_OP_FAIL_Z_KEEP
| BGFX_STENCIL_OP_PASS_Z_INCR
, 0
| BGFX_STENCIL_TEST_ALWAYS
| BGFX_STENCIL_FUNC_RMASK(0xff)
| BGFX_STENCIL_OP_FAIL_S_KEEP
| BGFX_STENCIL_OP_FAIL_Z_KEEP
| BGFX_STENCIL_OP_PASS_Z_DECR
);
bgfx::setVertexBuffer(&gl->tvb);
bgfx::setTexture(0, gl->s_tex, gl->th);
fan(paths[i].fillOffset, paths[i].fillCount);
bgfx::submit(gl->m_viewId, gl->prog);
}
}
// Draw aliased off-pixels
nvgRenderSetUniforms(gl, call->uniformOffset + gl->fragSize, call->image);
if (gl->edgeAntiAlias)
{
// Draw fringes
for (i = 0; i < npaths; i++)
{
bgfx::setState(gl->state
| BGFX_STATE_PT_TRISTRIP
);
bgfx::setStencil(0
| BGFX_STENCIL_TEST_EQUAL
| BGFX_STENCIL_FUNC_RMASK(0xff)
| BGFX_STENCIL_OP_FAIL_S_KEEP
| BGFX_STENCIL_OP_FAIL_Z_KEEP
| BGFX_STENCIL_OP_PASS_Z_KEEP
);
bgfx::setVertexBuffer(&gl->tvb, paths[i].strokeOffset, paths[i].strokeCount);
bgfx::setTexture(0, gl->s_tex, gl->th);
bgfx::submit(gl->m_viewId, gl->prog);
}
}
// Draw fill
bgfx::setState(gl->state);
bgfx::setVertexBuffer(&gl->tvb, call->vertexOffset, call->vertexCount);
bgfx::setTexture(0, gl->s_tex, gl->th);
bgfx::setStencil(0
| BGFX_STENCIL_TEST_NOTEQUAL
| BGFX_STENCIL_FUNC_RMASK(0xff)
| BGFX_STENCIL_OP_FAIL_S_ZERO
| BGFX_STENCIL_OP_FAIL_Z_ZERO
| BGFX_STENCIL_OP_PASS_Z_ZERO
);
bgfx::submit(gl->m_viewId, gl->prog);
}
static void glnvg__convexFill(struct GLNVGcontext* gl, struct GLNVGcall* call)
{
struct GLNVGpath* paths = &gl->paths[call->pathOffset];
int i, npaths = call->pathCount;
nvgRenderSetUniforms(gl, call->uniformOffset, call->image);
for (i = 0; i < npaths; i++)
{
if (paths[i].fillCount == 0) continue;
bgfx::setState(gl->state);
bgfx::setVertexBuffer(&gl->tvb);
bgfx::setTexture(0, gl->s_tex, gl->th);
fan(paths[i].fillOffset, paths[i].fillCount);
bgfx::submit(gl->m_viewId, gl->prog);
}
if (gl->edgeAntiAlias)
{
// Draw fringes
for (i = 0; i < npaths; i++)
{
bgfx::setState(gl->state
| BGFX_STATE_PT_TRISTRIP
);
bgfx::setVertexBuffer(&gl->tvb, paths[i].strokeOffset, paths[i].strokeCount);
bgfx::setTexture(0, gl->s_tex, gl->th);
bgfx::submit(gl->m_viewId, gl->prog);
}
}
}
static void glnvg__stroke(struct GLNVGcontext* gl, struct GLNVGcall* call)
{
struct GLNVGpath* paths = &gl->paths[call->pathOffset];
int npaths = call->pathCount, i;
nvgRenderSetUniforms(gl, call->uniformOffset, call->image);
// Draw Strokes
for (i = 0; i < npaths; i++)
{
bgfx::setState(gl->state
| BGFX_STATE_PT_TRISTRIP
);
bgfx::setVertexBuffer(&gl->tvb, paths[i].strokeOffset, paths[i].strokeCount);
bgfx::setTexture(0, gl->s_tex, gl->th);
bgfx::submit(gl->m_viewId, gl->prog);
}
}
static void glnvg__triangles(struct GLNVGcontext* gl, struct GLNVGcall* call)
{
if (3 <= call->vertexCount)
{
nvgRenderSetUniforms(gl, call->uniformOffset, call->image);
bgfx::setState(gl->state);
bgfx::setVertexBuffer(&gl->tvb, call->vertexOffset, call->vertexCount);
bgfx::setTexture(0, gl->s_tex, gl->th);
bgfx::submit(gl->m_viewId, gl->prog);
}
}
static const uint64_t s_blend[] =
{
BGFX_STATE_BLEND_ZERO,
BGFX_STATE_BLEND_ONE,
BGFX_STATE_BLEND_SRC_COLOR,
BGFX_STATE_BLEND_INV_SRC_COLOR,
BGFX_STATE_BLEND_DST_COLOR,
BGFX_STATE_BLEND_INV_DST_COLOR,
BGFX_STATE_BLEND_SRC_ALPHA,
BGFX_STATE_BLEND_INV_SRC_ALPHA,
BGFX_STATE_BLEND_DST_ALPHA,
BGFX_STATE_BLEND_INV_DST_ALPHA,
BGFX_STATE_BLEND_SRC_ALPHA_SAT,
};
static uint64_t glnvg_convertBlendFuncFactor(int factor)
{
const uint32_t numtz = bx::uint32_cnttz(factor);
const uint32_t idx = bx::uint32_min(numtz, BX_COUNTOF(s_blend)-1);
return s_blend[idx];
}
2017-04-11 08:16:10 +02:00
static GLNVGblend glnvg__blendCompositeOperation(NVGcompositeOperationState op)
2016-08-29 22:31:11 +02:00
{
2017-04-11 08:16:10 +02:00
GLNVGblend blend;
blend.srcRGB = glnvg_convertBlendFuncFactor(op.srcRGB);
blend.dstRGB = glnvg_convertBlendFuncFactor(op.dstRGB);
blend.srcAlpha = glnvg_convertBlendFuncFactor(op.srcAlpha);
blend.dstAlpha = glnvg_convertBlendFuncFactor(op.dstAlpha);
if (blend.srcRGB == BGFX_STATE_NONE || blend.dstRGB == BGFX_STATE_NONE || blend.srcAlpha == BGFX_STATE_NONE || blend.dstAlpha == BGFX_STATE_NONE)
{
blend.srcRGB = BGFX_STATE_BLEND_ONE;
blend.dstRGB = BGFX_STATE_BLEND_INV_SRC_ALPHA;
blend.srcAlpha = BGFX_STATE_BLEND_ONE;
blend.dstAlpha = BGFX_STATE_BLEND_INV_SRC_ALPHA;
}
return blend;
2016-08-29 22:31:11 +02:00
}
2017-04-11 08:16:10 +02:00
static void nvgRenderFlush(void* _userPtr)
2016-08-29 22:31:11 +02:00
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
if (gl->ncalls > 0)
{
bgfx::allocTransientVertexBuffer(&gl->tvb, gl->nverts, s_nvgDecl);
int allocated = gl->tvb.size/gl->tvb.stride;
if (allocated < gl->nverts)
{
gl->nverts = allocated;
BX_WARN(true, "Vertex number truncated due to transient vertex buffer overflow");
}
2017-04-11 08:16:10 +02:00
bx::memCopy(gl->tvb.data, gl->verts, gl->nverts * sizeof(struct NVGvertex) );
2016-08-29 22:31:11 +02:00
bgfx::setUniform(gl->u_viewSize, gl->view);
for (uint32_t ii = 0, num = gl->ncalls; ii < num; ++ii)
{
struct GLNVGcall* call = &gl->calls[ii];
2017-04-11 08:16:10 +02:00
const GLNVGblend* blend = &call->blendFunc;
gl->state = BGFX_STATE_BLEND_FUNC_SEPARATE(blend->srcRGB, blend->dstRGB, blend->srcAlpha, blend->dstAlpha)
| BGFX_STATE_RGB_WRITE
| BGFX_STATE_ALPHA_WRITE
;
2016-08-29 22:31:11 +02:00
switch (call->type)
{
case GLNVG_FILL:
glnvg__fill(gl, call);
break;
case GLNVG_CONVEXFILL:
glnvg__convexFill(gl, call);
break;
case GLNVG_STROKE:
glnvg__stroke(gl, call);
break;
case GLNVG_TRIANGLES:
glnvg__triangles(gl, call);
break;
}
}
}
// Reset calls
gl->nverts = 0;
gl->npaths = 0;
gl->ncalls = 0;
gl->nuniforms = 0;
}
static int glnvg__maxVertCount(const struct NVGpath* paths, int npaths)
{
int i, count = 0;
for (i = 0; i < npaths; i++)
{
count += paths[i].nfill;
count += paths[i].nstroke;
}
return count;
}
static int glnvg__maxi(int a, int b) { return a > b ? a : b; }
static struct GLNVGcall* glnvg__allocCall(struct GLNVGcontext* gl)
{
struct GLNVGcall* ret = NULL;
if (gl->ncalls+1 > gl->ccalls)
{
gl->ccalls = gl->ccalls == 0 ? 32 : gl->ccalls * 2;
gl->calls = (struct GLNVGcall*)BX_REALLOC(gl->m_allocator, gl->calls, sizeof(struct GLNVGcall) * gl->ccalls);
}
ret = &gl->calls[gl->ncalls++];
2017-04-11 08:16:10 +02:00
bx::memSet(ret, 0, sizeof(struct GLNVGcall) );
2016-08-29 22:31:11 +02:00
return ret;
}
static int glnvg__allocPaths(struct GLNVGcontext* gl, int n)
{
int ret = 0;
if (gl->npaths + n > gl->cpaths) {
GLNVGpath* paths;
int cpaths = glnvg__maxi(gl->npaths + n, 128) + gl->cpaths / 2; // 1.5x Overallocate
paths = (GLNVGpath*)BX_REALLOC(gl->m_allocator, gl->paths, sizeof(GLNVGpath) * cpaths);
if (paths == NULL) return -1;
gl->paths = paths;
gl->cpaths = cpaths;
}
ret = gl->npaths;
gl->npaths += n;
return ret;
}
static int glnvg__allocVerts(GLNVGcontext* gl, int n)
{
int ret = 0;
if (gl->nverts+n > gl->cverts)
{
NVGvertex* verts;
int cverts = glnvg__maxi(gl->nverts + n, 4096) + gl->cverts/2; // 1.5x Overallocate
verts = (NVGvertex*)BX_REALLOC(gl->m_allocator, gl->verts, sizeof(NVGvertex) * cverts);
if (verts == NULL) return -1;
gl->verts = verts;
gl->cverts = cverts;
}
ret = gl->nverts;
gl->nverts += n;
return ret;
}
static int glnvg__allocFragUniforms(struct GLNVGcontext* gl, int n)
{
int ret = 0, structSize = gl->fragSize;
if (gl->nuniforms+n > gl->cuniforms)
{
gl->cuniforms = gl->cuniforms == 0 ? glnvg__maxi(n, 32) : gl->cuniforms * 2;
gl->uniforms = (unsigned char*)BX_REALLOC(gl->m_allocator, gl->uniforms, gl->cuniforms * structSize);
}
ret = gl->nuniforms * structSize;
gl->nuniforms += n;
return ret;
}
static void glnvg__vset(struct NVGvertex* vtx, float x, float y, float u, float v)
{
vtx->x = x;
vtx->y = y;
vtx->u = u;
vtx->v = v;
}
2017-04-11 08:16:10 +02:00
static void nvgRenderFill(void* _userPtr, NVGpaint* paint, NVGcompositeOperationState compositeOperation, NVGscissor* scissor,
float fringe, const float* bounds, const NVGpath* paths, int npaths)
2016-08-29 22:31:11 +02:00
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
struct GLNVGcall* call = glnvg__allocCall(gl);
struct NVGvertex* quad;
struct GLNVGfragUniforms* frag;
int i, maxverts, offset;
call->type = GLNVG_FILL;
call->pathOffset = glnvg__allocPaths(gl, npaths);
call->pathCount = npaths;
call->image = paint->image;
2017-04-11 08:16:10 +02:00
call->blendFunc = glnvg__blendCompositeOperation(compositeOperation);
2016-08-29 22:31:11 +02:00
if (npaths == 1 && paths[0].convex)
{
call->type = GLNVG_CONVEXFILL;
}
// Allocate vertices for all the paths.
maxverts = glnvg__maxVertCount(paths, npaths) + 6;
offset = glnvg__allocVerts(gl, maxverts);
for (i = 0; i < npaths; i++)
{
struct GLNVGpath* copy = &gl->paths[call->pathOffset + i];
const struct NVGpath* path = &paths[i];
2017-04-11 08:16:10 +02:00
bx::memSet(copy, 0, sizeof(struct GLNVGpath) );
2016-08-29 22:31:11 +02:00
if (path->nfill > 0)
{
copy->fillOffset = offset;
copy->fillCount = path->nfill;
2017-04-11 08:16:10 +02:00
bx::memCopy(&gl->verts[offset], path->fill, sizeof(struct NVGvertex) * path->nfill);
2016-08-29 22:31:11 +02:00
offset += path->nfill;
}
if (path->nstroke > 0)
{
copy->strokeOffset = offset;
copy->strokeCount = path->nstroke;
2017-04-11 08:16:10 +02:00
bx::memCopy(&gl->verts[offset], path->stroke, sizeof(struct NVGvertex) * path->nstroke);
2016-08-29 22:31:11 +02:00
offset += path->nstroke;
}
}
// Quad
call->vertexOffset = offset;
call->vertexCount = 6;
quad = &gl->verts[call->vertexOffset];
glnvg__vset(&quad[0], bounds[0], bounds[3], 0.5f, 1.0f);
glnvg__vset(&quad[1], bounds[2], bounds[3], 0.5f, 1.0f);
glnvg__vset(&quad[2], bounds[2], bounds[1], 0.5f, 1.0f);
glnvg__vset(&quad[3], bounds[0], bounds[3], 0.5f, 1.0f);
glnvg__vset(&quad[4], bounds[2], bounds[1], 0.5f, 1.0f);
glnvg__vset(&quad[5], bounds[0], bounds[1], 0.5f, 1.0f);
// Setup uniforms for draw calls
if (call->type == GLNVG_FILL)
{
call->uniformOffset = glnvg__allocFragUniforms(gl, 2);
// Simple shader for stencil
frag = nvg__fragUniformPtr(gl, call->uniformOffset);
2017-04-11 08:16:10 +02:00
bx::memSet(frag, 0, sizeof(*frag) );
2016-08-29 22:31:11 +02:00
frag->type = NSVG_SHADER_SIMPLE;
// Fill shader
glnvg__convertPaint(gl, nvg__fragUniformPtr(gl, call->uniformOffset + gl->fragSize), paint, scissor, fringe, fringe);
}
else
{
call->uniformOffset = glnvg__allocFragUniforms(gl, 1);
// Fill shader
glnvg__convertPaint(gl, nvg__fragUniformPtr(gl, call->uniformOffset), paint, scissor, fringe, fringe);
}
}
2017-04-11 08:16:10 +02:00
static void nvgRenderStroke(void* _userPtr, struct NVGpaint* paint, NVGcompositeOperationState compositeOperation, struct NVGscissor* scissor, float fringe,
2016-08-29 22:31:11 +02:00
float strokeWidth, const struct NVGpath* paths, int npaths)
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
struct GLNVGcall* call = glnvg__allocCall(gl);
int i, maxverts, offset;
call->type = GLNVG_STROKE;
call->pathOffset = glnvg__allocPaths(gl, npaths);
call->pathCount = npaths;
call->image = paint->image;
2017-04-11 08:16:10 +02:00
call->blendFunc = glnvg__blendCompositeOperation(compositeOperation);
2016-08-29 22:31:11 +02:00
// Allocate vertices for all the paths.
maxverts = glnvg__maxVertCount(paths, npaths);
offset = glnvg__allocVerts(gl, maxverts);
for (i = 0; i < npaths; i++)
{
struct GLNVGpath* copy = &gl->paths[call->pathOffset + i];
const struct NVGpath* path = &paths[i];
2017-04-11 08:16:10 +02:00
bx::memSet(copy, 0, sizeof(struct GLNVGpath) );
2016-08-29 22:31:11 +02:00
if (path->nstroke)
{
copy->strokeOffset = offset;
copy->strokeCount = path->nstroke;
2017-04-11 08:16:10 +02:00
bx::memCopy(&gl->verts[offset], path->stroke, sizeof(struct NVGvertex) * path->nstroke);
2016-08-29 22:31:11 +02:00
offset += path->nstroke;
}
}
// Fill shader
call->uniformOffset = glnvg__allocFragUniforms(gl, 1);
glnvg__convertPaint(gl, nvg__fragUniformPtr(gl, call->uniformOffset), paint, scissor, strokeWidth, fringe);
}
2017-04-11 08:16:10 +02:00
static void nvgRenderTriangles(void* _userPtr, struct NVGpaint* paint, NVGcompositeOperationState compositeOperation, struct NVGscissor* scissor,
2016-08-29 22:31:11 +02:00
const struct NVGvertex* verts, int nverts)
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
struct GLNVGcall* call = glnvg__allocCall(gl);
struct GLNVGfragUniforms* frag;
call->type = GLNVG_TRIANGLES;
call->image = paint->image;
2017-04-11 08:16:10 +02:00
call->blendFunc = glnvg__blendCompositeOperation(compositeOperation);
2016-08-29 22:31:11 +02:00
// Allocate vertices for all the paths.
call->vertexOffset = glnvg__allocVerts(gl, nverts);
call->vertexCount = nverts;
2017-04-11 08:16:10 +02:00
bx::memCopy(&gl->verts[call->vertexOffset], verts, sizeof(struct NVGvertex) * nverts);
2016-08-29 22:31:11 +02:00
// Fill shader
call->uniformOffset = glnvg__allocFragUniforms(gl, 1);
frag = nvg__fragUniformPtr(gl, call->uniformOffset);
glnvg__convertPaint(gl, frag, paint, scissor, 1.0f, 1.0f);
frag->type = NSVG_SHADER_IMG;
}
static void nvgRenderDelete(void* _userPtr)
{
struct GLNVGcontext* gl = (struct GLNVGcontext*)_userPtr;
if (gl == NULL)
{
return;
}
bgfx::destroyProgram(gl->prog);
bgfx::destroyTexture(gl->texMissing);
bgfx::destroyUniform(gl->u_scissorMat);
bgfx::destroyUniform(gl->u_paintMat);
bgfx::destroyUniform(gl->u_innerCol);
bgfx::destroyUniform(gl->u_outerCol);
bgfx::destroyUniform(gl->u_viewSize);
bgfx::destroyUniform(gl->u_scissorExtScale);
bgfx::destroyUniform(gl->u_extentRadius);
bgfx::destroyUniform(gl->u_params);
bgfx::destroyUniform(gl->s_tex);
if (bgfx::isValid(gl->u_halfTexel) )
{
bgfx::destroyUniform(gl->u_halfTexel);
}
for (uint32_t ii = 0, num = gl->ntextures; ii < num; ++ii)
{
if (bgfx::isValid(gl->textures[ii].id)
&& (gl->textures[ii].flags & NVG_IMAGE_NODELETE) == 0)
{
bgfx::destroyTexture(gl->textures[ii].id);
}
}
BX_FREE(gl->m_allocator, gl->uniforms);
BX_FREE(gl->m_allocator, gl->verts);
BX_FREE(gl->m_allocator, gl->paths);
BX_FREE(gl->m_allocator, gl->calls);
BX_FREE(gl->m_allocator, gl->textures);
BX_FREE(gl->m_allocator, gl);
}
} // namespace
NVGcontext* nvgCreate(int edgeaa, unsigned char _viewId, bx::AllocatorI* _allocator)
{
if (NULL == _allocator)
{
#if BX_CONFIG_ALLOCATOR_CRT
static bx::CrtAllocator allocator;
_allocator = &allocator;
#else
BX_CHECK(false, "No allocator has been passed to nvgCreate(). Either specify a bx::AllocatorI instance or enable BX_CONFIG_ALLOCATOR_CRT directive.");
return NULL;
#endif // BX_CONFIG_ALLOCATOR_CRT
}
struct NVGparams params;
struct NVGcontext* ctx = NULL;
struct GLNVGcontext* gl = (struct GLNVGcontext*)BX_ALLOC(_allocator, sizeof(struct GLNVGcontext) );
if (gl == NULL) goto error;
2017-04-11 08:16:10 +02:00
bx::memSet(gl, 0, sizeof(struct GLNVGcontext) );
2016-08-29 22:31:11 +02:00
2017-04-11 08:16:10 +02:00
bx::memSet(&params, 0, sizeof(params) );
2016-08-29 22:31:11 +02:00
params.renderCreate = nvgRenderCreate;
params.renderCreateTexture = nvgRenderCreateTexture;
params.renderDeleteTexture = nvgRenderDeleteTexture;
params.renderUpdateTexture = nvgRenderUpdateTexture;
params.renderGetTextureSize = nvgRenderGetTextureSize;
params.renderViewport = nvgRenderViewport;
params.renderFlush = nvgRenderFlush;
params.renderFill = nvgRenderFill;
params.renderStroke = nvgRenderStroke;
params.renderTriangles = nvgRenderTriangles;
params.renderDelete = nvgRenderDelete;
params.userPtr = gl;
params.edgeAntiAlias = edgeaa;
gl->m_allocator = _allocator;
gl->edgeAntiAlias = edgeaa;
gl->m_viewId = uint8_t(_viewId);
ctx = nvgCreateInternal(&params);
if (ctx == NULL) goto error;
return ctx;
error:
// 'gl' is freed by nvgDeleteInternal.
if (ctx != NULL)
{
nvgDeleteInternal(ctx);
}
return NULL;
}
NVGcontext* nvgCreate(int edgeaa, unsigned char _viewId) {
return nvgCreate(edgeaa, _viewId, NULL);
}
void nvgDelete(struct NVGcontext* ctx)
{
nvgDeleteInternal(ctx);
}
uint8_t nvgViewId(struct NVGcontext* ctx)
{
struct NVGparams* params = nvgInternalParams(ctx);
struct GLNVGcontext* gl = (struct GLNVGcontext*)params->userPtr;
return gl->m_viewId;
}
void nvgViewId(struct NVGcontext* ctx, unsigned char _viewId)
{
struct NVGparams* params = nvgInternalParams(ctx);
struct GLNVGcontext* gl = (struct GLNVGcontext*)params->userPtr;
gl->m_viewId = uint8_t(_viewId);
}
bgfx::TextureHandle nvglImageHandle(NVGcontext* ctx, int image)
{
GLNVGcontext* gl = (GLNVGcontext*)nvgInternalParams(ctx)->userPtr;
GLNVGtexture* tex = glnvg__findTexture(gl, image);
return tex->id;
}
2017-04-11 08:16:10 +02:00
NVGLUframebuffer* nvgluCreateFramebuffer(NVGcontext* ctx, int width, int height, int imageFlags, uint8_t viewId)
{
NVGLUframebuffer* framebuffer = nvgluCreateFramebuffer(ctx, width, height, imageFlags);
if (framebuffer != NULL)
{
nvgluSetViewFramebuffer(viewId, framebuffer);
}
return framebuffer;
}
NVGLUframebuffer* nvgluCreateFramebuffer(NVGcontext* ctx, int width, int height, int imageFlags)
{
2016-08-29 22:31:11 +02:00
NVGLUframebuffer* framebuffer = new NVGLUframebuffer;
framebuffer->ctx = ctx;
framebuffer->image = nvgCreateImageRGBA(ctx, width, height, imageFlags | NVG_IMAGE_PREMULTIPLIED, NULL);
bgfx::TextureHandle texture = nvglImageHandle(ctx, framebuffer->image);
2017-04-11 08:16:10 +02:00
if (!bgfx::isValid(texture))
{
2016-08-29 22:31:11 +02:00
nvgluDeleteFramebuffer(framebuffer);
return NULL;
}
framebuffer->handle = bgfx::createFrameBuffer(1, &texture, false);
if (!bgfx::isValid(framebuffer->handle))
{
nvgluDeleteFramebuffer(framebuffer);
return NULL;
}
return framebuffer;
}
2017-04-11 08:16:10 +02:00
void nvgluBindFramebuffer(NVGLUframebuffer* framebuffer)
{
2016-08-29 22:31:11 +02:00
static NVGcontext* s_prevCtx = NULL;
static uint8_t s_prevViewId;
2017-04-11 08:16:10 +02:00
if (framebuffer != NULL)
{
2016-08-29 22:31:11 +02:00
s_prevCtx = framebuffer->ctx;
s_prevViewId = nvgViewId(framebuffer->ctx);
nvgViewId(framebuffer->ctx, framebuffer->viewId);
} else if (s_prevCtx != NULL) {
nvgViewId(s_prevCtx, s_prevViewId);
}
}
2017-04-11 08:16:10 +02:00
void nvgluDeleteFramebuffer(NVGLUframebuffer* framebuffer)
{
2016-08-29 22:31:11 +02:00
if (framebuffer == NULL)
return;
if (bgfx::isValid(framebuffer->handle))
2017-04-11 08:16:10 +02:00
{
2016-08-29 22:31:11 +02:00
bgfx::destroyFrameBuffer(framebuffer->handle);
2017-04-11 08:16:10 +02:00
}
2016-08-29 22:31:11 +02:00
if (framebuffer->image > 0)
2017-04-11 08:16:10 +02:00
{
2016-08-29 22:31:11 +02:00
nvgDeleteImage(framebuffer->ctx, framebuffer->image);
2017-04-11 08:16:10 +02:00
}
2016-08-29 22:31:11 +02:00
delete framebuffer;
}
2017-04-11 08:16:10 +02:00
void nvgluSetViewFramebuffer(uint8_t viewId, NVGLUframebuffer* framebuffer)
{
framebuffer->viewId = viewId;
bgfx::setViewFrameBuffer(viewId, framebuffer->handle);
bgfx::setViewSeq(viewId, true);
}