| 1 | /* | 
| 2 |  * Copyright 2018 Google Inc. | 
| 3 |  * | 
| 4 |  * Use of this source code is governed by a BSD-style license that can be | 
| 5 |  * found in the LICENSE file. | 
| 6 |  */ | 
| 7 |  | 
| 8 | #include "src/gpu/ccpr/GrCCPerFlushResources.h" | 
| 9 |  | 
| 10 | #include "include/gpu/GrRecordingContext.h" | 
| 11 | #include "src/gpu/GrMemoryPool.h" | 
| 12 | #include "src/gpu/GrOnFlushResourceProvider.h" | 
| 13 | #include "src/gpu/GrRecordingContextPriv.h" | 
| 14 | #include "src/gpu/GrRenderTargetContext.h" | 
| 15 | #include "src/gpu/GrSurfaceContextPriv.h" | 
| 16 | #include "src/gpu/ccpr/GrCCPathCache.h" | 
| 17 | #include "src/gpu/ccpr/GrGSCoverageProcessor.h" | 
| 18 | #include "src/gpu/ccpr/GrSampleMaskProcessor.h" | 
| 19 | #include "src/gpu/ccpr/GrVSCoverageProcessor.h" | 
| 20 | #include "src/gpu/geometry/GrStyledShape.h" | 
| 21 | #include <algorithm> | 
| 22 |  | 
| 23 | using CoverageType = GrCCAtlas::CoverageType; | 
| 24 | using FillBatchID = GrCCFiller::BatchID; | 
| 25 | using StrokeBatchID = GrCCStroker::BatchID; | 
| 26 | using PathInstance = GrCCPathProcessor::Instance; | 
| 27 |  | 
| 28 | static constexpr int kFillIdx = GrCCPerFlushResourceSpecs::kFillIdx; | 
| 29 | static constexpr int kStrokeIdx = GrCCPerFlushResourceSpecs::kStrokeIdx; | 
| 30 |  | 
| 31 | namespace { | 
| 32 |  | 
| 33 | // Base class for an Op that renders a CCPR atlas. | 
| 34 | class AtlasOp : public GrDrawOp { | 
| 35 | public: | 
| 36 |     FixedFunctionFlags fixedFunctionFlags() const override { return FixedFunctionFlags::kNone; } | 
| 37 |     GrProcessorSet::Analysis finalize(const GrCaps&, const GrAppliedClip*, | 
| 38 |                                       bool hasMixedSampledCoverage, GrClampType) override { | 
| 39 |         return GrProcessorSet::EmptySetAnalysis(); | 
| 40 |     } | 
| 41 |     CombineResult onCombineIfPossible(GrOp* other, GrRecordingContext::Arenas*, | 
| 42 |                                       const GrCaps&) override { | 
| 43 |         // We will only make multiple copy ops if they have different source proxies. | 
| 44 |         // TODO: make use of texture chaining. | 
| 45 |         return CombineResult::kCannotCombine; | 
| 46 |     } | 
| 47 |  | 
| 48 | protected: | 
| 49 |     AtlasOp(uint32_t classID, sk_sp<const GrCCPerFlushResources> resources, | 
| 50 |             const SkISize& drawBounds) | 
| 51 |             : GrDrawOp(classID) | 
| 52 |             , fResources(std::move(resources)) { | 
| 53 |         this->setBounds(SkRect::MakeIWH(drawBounds.width(), drawBounds.height()), | 
| 54 |                         GrOp::HasAABloat::kNo, GrOp::IsHairline::kNo); | 
| 55 |     } | 
| 56 |  | 
| 57 |     const sk_sp<const GrCCPerFlushResources> fResources; | 
| 58 |  | 
| 59 | private: | 
| 60 |     void onPrePrepare(GrRecordingContext*, | 
| 61 |                       const GrSurfaceProxyView* writeView, | 
| 62 |                       GrAppliedClip*, | 
| 63 |                       const GrXferProcessor::DstProxyView&) final {} | 
| 64 |     void onPrepare(GrOpFlushState*) final {} | 
| 65 | }; | 
| 66 |  | 
| 67 | // Copies paths from a cached coverage count or msaa atlas into an 8-bit literal-coverage atlas. | 
| 68 | class CopyAtlasOp : public AtlasOp { | 
| 69 | public: | 
| 70 |     DEFINE_OP_CLASS_ID | 
| 71 |  | 
| 72 |     static std::unique_ptr<GrDrawOp> Make( | 
| 73 |             GrRecordingContext* context, sk_sp<const GrCCPerFlushResources> resources, | 
| 74 |             sk_sp<GrTextureProxy> copyProxy, int baseInstance, int endInstance, | 
| 75 |             const SkISize& drawBounds) { | 
| 76 |         GrOpMemoryPool* pool = context->priv().opMemoryPool(); | 
| 77 |  | 
| 78 |         return pool->allocate<CopyAtlasOp>(std::move(resources), std::move(copyProxy), baseInstance, | 
| 79 |                                            endInstance, drawBounds); | 
| 80 |     } | 
| 81 |  | 
| 82 |     const char* name() const override { return "CopyAtlasOp (CCPR)" ; } | 
| 83 |  | 
| 84 |     void visitProxies(const VisitProxyFunc& fn) const override { | 
| 85 |         fn(fSrcProxy.get(), GrMipmapped::kNo); | 
| 86 |     } | 
| 87 |  | 
| 88 |     void onExecute(GrOpFlushState* flushState, const SkRect& chainBounds) override { | 
| 89 |         SkASSERT(fSrcProxy); | 
| 90 |         SkASSERT(fSrcProxy->isInstantiated()); | 
| 91 |  | 
| 92 |         auto coverageMode = GrCCAtlas::CoverageTypeToPathCoverageMode( | 
| 93 |                 fResources->renderedPathCoverageType()); | 
| 94 |         GrColorType ct = GrCCAtlas::CoverageTypeToColorType(fResources->renderedPathCoverageType()); | 
| 95 |         GrSwizzle swizzle = flushState->caps().getReadSwizzle(fSrcProxy->backendFormat(), ct); | 
| 96 |         GrCCPathProcessor pathProc(coverageMode, fSrcProxy->peekTexture(), swizzle, | 
| 97 |                                    GrCCAtlas::kTextureOrigin); | 
| 98 |  | 
| 99 |         bool hasScissor = flushState->appliedClip() && | 
| 100 |                           flushState->appliedClip()->scissorState().enabled(); | 
| 101 |         GrPipeline pipeline(hasScissor ? GrScissorTest::kEnabled : GrScissorTest::kDisabled, | 
| 102 |                             SkBlendMode::kSrc, flushState->drawOpArgs().writeSwizzle()); | 
| 103 |  | 
| 104 |         pathProc.drawPaths(flushState, pipeline, *fSrcProxy, *fResources, fBaseInstance, | 
| 105 |                            fEndInstance, this->bounds()); | 
| 106 |     } | 
| 107 |  | 
| 108 | private: | 
| 109 |     friend class ::GrOpMemoryPool; // for ctor | 
| 110 |  | 
| 111 |     CopyAtlasOp(sk_sp<const GrCCPerFlushResources> resources, sk_sp<GrTextureProxy> srcProxy, | 
| 112 |                 int baseInstance, int endInstance, const SkISize& drawBounds) | 
| 113 |             : AtlasOp(ClassID(), std::move(resources), drawBounds) | 
| 114 |             , fSrcProxy(srcProxy) | 
| 115 |             , fBaseInstance(baseInstance) | 
| 116 |             , fEndInstance(endInstance) { | 
| 117 |     } | 
| 118 |     sk_sp<GrTextureProxy> fSrcProxy; | 
| 119 |     const int fBaseInstance; | 
| 120 |     const int fEndInstance; | 
| 121 | }; | 
| 122 |  | 
| 123 | // Renders coverage counts to a CCPR atlas using the resources' pre-filled GrCCPathParser. | 
| 124 | template<typename ProcessorType> class RenderAtlasOp : public AtlasOp { | 
| 125 | public: | 
| 126 |     DEFINE_OP_CLASS_ID | 
| 127 |  | 
| 128 |     static std::unique_ptr<GrDrawOp> Make( | 
| 129 |             GrRecordingContext* context, sk_sp<const GrCCPerFlushResources> resources, | 
| 130 |             FillBatchID fillBatchID, StrokeBatchID strokeBatchID, const SkISize& drawBounds) { | 
| 131 |         GrOpMemoryPool* pool = context->priv().opMemoryPool(); | 
| 132 |  | 
| 133 |         return pool->allocate<RenderAtlasOp>( | 
| 134 |                 std::move(resources), fillBatchID, strokeBatchID, drawBounds); | 
| 135 |     } | 
| 136 |  | 
| 137 |     // GrDrawOp interface. | 
| 138 |     const char* name() const override { return "RenderAtlasOp (CCPR)" ; } | 
| 139 |  | 
| 140 |     void onExecute(GrOpFlushState* flushState, const SkRect& chainBounds) override { | 
| 141 |         ProcessorType proc; | 
| 142 |         GrPipeline pipeline(GrScissorTest::kEnabled, SkBlendMode::kPlus, | 
| 143 |                             flushState->drawOpArgs().writeSwizzle()); | 
| 144 |         fResources->filler().drawFills(flushState, &proc, pipeline, fFillBatchID, fDrawBounds); | 
| 145 |         fResources->stroker().drawStrokes(flushState, &proc, fStrokeBatchID, fDrawBounds); | 
| 146 |     } | 
| 147 |  | 
| 148 | private: | 
| 149 |     friend class ::GrOpMemoryPool; // for ctor | 
| 150 |  | 
| 151 |     RenderAtlasOp(sk_sp<const GrCCPerFlushResources> resources, FillBatchID fillBatchID, | 
| 152 |                   StrokeBatchID strokeBatchID, const SkISize& drawBounds) | 
| 153 |             : AtlasOp(ClassID(), std::move(resources), drawBounds) | 
| 154 |             , fFillBatchID(fillBatchID) | 
| 155 |             , fStrokeBatchID(strokeBatchID) | 
| 156 |             , fDrawBounds(SkIRect::MakeWH(drawBounds.width(), drawBounds.height())) { | 
| 157 |     } | 
| 158 |  | 
| 159 |     const FillBatchID fFillBatchID; | 
| 160 |     const StrokeBatchID fStrokeBatchID; | 
| 161 |     const SkIRect fDrawBounds; | 
| 162 | }; | 
| 163 |  | 
| 164 | }  // namespace | 
| 165 |  | 
| 166 | static int inst_buffer_count(const GrCCPerFlushResourceSpecs& specs) { | 
| 167 |     return specs.fNumCachedPaths + | 
| 168 |            // Copies get two instances per draw: 1 copy + 1 draw. | 
| 169 |            (specs.fNumCopiedPaths[kFillIdx] + specs.fNumCopiedPaths[kStrokeIdx]) * 2 + | 
| 170 |            specs.fNumRenderedPaths[kFillIdx] + specs.fNumRenderedPaths[kStrokeIdx]; | 
| 171 |            // No clips in instance buffers. | 
| 172 | } | 
| 173 |  | 
| 174 | GrCCPerFlushResources::GrCCPerFlushResources( | 
| 175 |         GrOnFlushResourceProvider* onFlushRP, CoverageType coverageType, | 
| 176 |         const GrCCPerFlushResourceSpecs& specs) | 
| 177 |         // Overallocate by one point so we can call Sk4f::Store at the final SkPoint in the array. | 
| 178 |         // (See transform_path_pts below.) | 
| 179 |         // FIXME: instead use built-in instructions to write only the first two lanes of an Sk4f. | 
| 180 |         : fLocalDevPtsBuffer(std::max(specs.fRenderedPathStats[kFillIdx].fMaxPointsPerPath, | 
| 181 |                                     specs.fRenderedPathStats[kStrokeIdx].fMaxPointsPerPath) + 1) | 
| 182 |         , fFiller((CoverageType::kFP16_CoverageCount == coverageType) | 
| 183 |                           ? GrCCFiller::Algorithm::kCoverageCount | 
| 184 |                           : GrCCFiller::Algorithm::kStencilWindingCount, | 
| 185 |                   specs.fNumRenderedPaths[kFillIdx] + specs.fNumClipPaths, | 
| 186 |                   specs.fRenderedPathStats[kFillIdx].fNumTotalSkPoints, | 
| 187 |                   specs.fRenderedPathStats[kFillIdx].fNumTotalSkVerbs, | 
| 188 |                   specs.fRenderedPathStats[kFillIdx].fNumTotalConicWeights) | 
| 189 |         , fStroker(specs.fNumRenderedPaths[kStrokeIdx], | 
| 190 |                    specs.fRenderedPathStats[kStrokeIdx].fNumTotalSkPoints, | 
| 191 |                    specs.fRenderedPathStats[kStrokeIdx].fNumTotalSkVerbs) | 
| 192 |         , fCopyAtlasStack(CoverageType::kA8_LiteralCoverage, specs.fCopyAtlasSpecs, | 
| 193 |                           onFlushRP->caps()) | 
| 194 |         , fRenderedAtlasStack(coverageType, specs.fRenderedAtlasSpecs, onFlushRP->caps()) | 
| 195 |         , fIndexBuffer(GrCCPathProcessor::FindIndexBuffer(onFlushRP)) | 
| 196 |         , fVertexBuffer(GrCCPathProcessor::FindVertexBuffer(onFlushRP)) | 
| 197 |         , fNextCopyInstanceIdx(0) | 
| 198 |         , fNextPathInstanceIdx( | 
| 199 |                 specs.fNumCopiedPaths[kFillIdx] + specs.fNumCopiedPaths[kStrokeIdx]) { | 
| 200 |     if (!fIndexBuffer) { | 
| 201 |         SkDebugf("WARNING: failed to allocate CCPR index buffer. No paths will be drawn.\n" ); | 
| 202 |         return; | 
| 203 |     } | 
| 204 |     if (!fVertexBuffer) { | 
| 205 |         SkDebugf("WARNING: failed to allocate CCPR vertex buffer. No paths will be drawn.\n" ); | 
| 206 |         return; | 
| 207 |     } | 
| 208 |     fPathInstanceBuffer.resetAndMapBuffer(onFlushRP, | 
| 209 |                                           inst_buffer_count(specs) * sizeof(PathInstance)); | 
| 210 |     if (!fPathInstanceBuffer.hasGpuBuffer()) { | 
| 211 |         SkDebugf("WARNING: failed to allocate CCPR instance buffer. No paths will be drawn.\n" ); | 
| 212 |         return; | 
| 213 |     } | 
| 214 |  | 
| 215 |     if (CoverageType::kA8_Multisample == coverageType) { | 
| 216 |         int numRenderedPaths = | 
| 217 |                 specs.fNumRenderedPaths[kFillIdx] + specs.fNumRenderedPaths[kStrokeIdx] + | 
| 218 |                 specs.fNumClipPaths; | 
| 219 |         fStencilResolveBuffer.resetAndMapBuffer( | 
| 220 |                 onFlushRP, numRenderedPaths * sizeof(GrStencilAtlasOp::ResolveRectInstance)); | 
| 221 |         if (!fStencilResolveBuffer.hasGpuBuffer()) { | 
| 222 |             SkDebugf("WARNING: failed to allocate CCPR stencil resolve buffer. "  | 
| 223 |                      "No paths will be drawn.\n" ); | 
| 224 |             return; | 
| 225 |         } | 
| 226 |         SkDEBUGCODE(fEndStencilResolveInstance = numRenderedPaths); | 
| 227 |     } | 
| 228 |  | 
| 229 |     SkDEBUGCODE(fEndCopyInstance = | 
| 230 |                         specs.fNumCopiedPaths[kFillIdx] + specs.fNumCopiedPaths[kStrokeIdx]); | 
| 231 |     SkDEBUGCODE(fEndPathInstance = inst_buffer_count(specs)); | 
| 232 | } | 
| 233 |  | 
| 234 | void GrCCPerFlushResources::upgradeEntryToLiteralCoverageAtlas( | 
| 235 |         GrCCPathCache* pathCache, GrOnFlushResourceProvider* onFlushRP, GrCCPathCacheEntry* entry, | 
| 236 |         GrFillRule fillRule) { | 
| 237 |     using ReleaseAtlasResult = GrCCPathCacheEntry::ReleaseAtlasResult; | 
| 238 |     SkASSERT(this->isMapped()); | 
| 239 |     SkASSERT(fNextCopyInstanceIdx < fEndCopyInstance); | 
| 240 |  | 
| 241 |     const GrCCCachedAtlas* cachedAtlas = entry->cachedAtlas(); | 
| 242 |     SkASSERT(cachedAtlas); | 
| 243 |     SkASSERT(cachedAtlas->getOnFlushProxy()); | 
| 244 |  | 
| 245 |     if (CoverageType::kA8_LiteralCoverage == cachedAtlas->coverageType()) { | 
| 246 |         // This entry has already been upgraded to literal coverage. The path must have been drawn | 
| 247 |         // multiple times during the flush. | 
| 248 |         SkDEBUGCODE(--fEndCopyInstance); | 
| 249 |         return; | 
| 250 |     } | 
| 251 |  | 
| 252 |     SkIVector newAtlasOffset; | 
| 253 |     if (GrCCAtlas* retiredAtlas = fCopyAtlasStack.addRect(entry->devIBounds(), &newAtlasOffset)) { | 
| 254 |         // We did not fit in the previous copy atlas and it was retired. We will render the ranges | 
| 255 |         // up until fCopyPathRanges.count() into the retired atlas during finalize(). | 
| 256 |         retiredAtlas->setFillBatchID(fCopyPathRanges.count()); | 
| 257 |         fCurrCopyAtlasRangesIdx = fCopyPathRanges.count(); | 
| 258 |     } | 
| 259 |  | 
| 260 |     this->recordCopyPathInstance( | 
| 261 |             *entry, newAtlasOffset, fillRule, sk_ref_sp(cachedAtlas->getOnFlushProxy())); | 
| 262 |  | 
| 263 |     sk_sp<GrTexture> previousAtlasTexture = | 
| 264 |             sk_ref_sp(cachedAtlas->getOnFlushProxy()->peekTexture()); | 
| 265 |     GrCCAtlas* newAtlas = &fCopyAtlasStack.current(); | 
| 266 |     if (ReleaseAtlasResult::kDidInvalidateFromCache == | 
| 267 |             entry->upgradeToLiteralCoverageAtlas(pathCache, onFlushRP, newAtlas, newAtlasOffset)) { | 
| 268 |         // This texture just got booted out of the cache. Keep it around, in case we might be able | 
| 269 |         // to recycle it for a new atlas. We can recycle it because copying happens before rendering | 
| 270 |         // new paths, and every path from the atlas that we're planning to use this flush will be | 
| 271 |         // copied to a new atlas. We'll never copy some and leave others. | 
| 272 |         fRecyclableAtlasTextures.push_back(std::move(previousAtlasTexture)); | 
| 273 |     } | 
| 274 | } | 
| 275 |  | 
| 276 | void GrCCPerFlushResources::recordCopyPathInstance( | 
| 277 |         const GrCCPathCacheEntry& entry, const SkIVector& newAtlasOffset, GrFillRule fillRule, | 
| 278 |         sk_sp<GrTextureProxy> srcProxy) { | 
| 279 |     SkASSERT(fNextCopyInstanceIdx < fEndCopyInstance); | 
| 280 |  | 
| 281 |     // Write the instance at the back of the array. | 
| 282 |     int currentInstanceIdx = fNextCopyInstanceIdx++; | 
| 283 |     fPathInstanceBuffer[currentInstanceIdx].set(entry, newAtlasOffset, SK_PMColor4fWHITE, fillRule); | 
| 284 |  | 
| 285 |     // Percolate the instance forward until it's contiguous with other instances that share the same | 
| 286 |     // proxy. | 
| 287 |     for (int i = fCopyPathRanges.count() - 1; i >= fCurrCopyAtlasRangesIdx; --i) { | 
| 288 |         if (fCopyPathRanges[i].fSrcProxy == srcProxy) { | 
| 289 |             ++fCopyPathRanges[i].fCount; | 
| 290 |             return; | 
| 291 |         } | 
| 292 |         int rangeFirstInstanceIdx = currentInstanceIdx - fCopyPathRanges[i].fCount; | 
| 293 |         std::swap(fPathInstanceBuffer[rangeFirstInstanceIdx], | 
| 294 |                   fPathInstanceBuffer[currentInstanceIdx]); | 
| 295 |         currentInstanceIdx = rangeFirstInstanceIdx; | 
| 296 |     } | 
| 297 |  | 
| 298 |     // An instance with this particular proxy did not yet exist in the array. Add a range for it, | 
| 299 |     // first moving any later ranges back to make space for it at fCurrCopyAtlasRangesIdx. | 
| 300 |     fCopyPathRanges.push_back(); | 
| 301 |     std::move_backward(fCopyPathRanges.begin() + fCurrCopyAtlasRangesIdx, | 
| 302 |                        fCopyPathRanges.end() - 1, | 
| 303 |                        fCopyPathRanges.end()); | 
| 304 |     fCopyPathRanges[fCurrCopyAtlasRangesIdx] = {std::move(srcProxy), 1}; | 
| 305 | } | 
| 306 |  | 
| 307 | static bool transform_path_pts( | 
| 308 |         const SkMatrix& m, const SkPath& path, const SkAutoSTArray<32, SkPoint>& outDevPts, | 
| 309 |         GrOctoBounds* octoBounds) { | 
| 310 |     const SkPoint* pts = SkPathPriv::PointData(path); | 
| 311 |     int numPts = path.countPoints(); | 
| 312 |     SkASSERT(numPts + 1 <= outDevPts.count()); | 
| 313 |     SkASSERT(numPts); | 
| 314 |  | 
| 315 |     // m45 transforms path points into "45 degree" device space. A bounding box in this space gives | 
| 316 |     // the circumscribing octagon's diagonals. We could use SK_ScalarRoot2Over2, but an orthonormal | 
| 317 |     // transform is not necessary as long as the shader uses the correct inverse. | 
| 318 |     SkMatrix m45; | 
| 319 |     m45.setSinCos(1, 1); | 
| 320 |     m45.preConcat(m); | 
| 321 |  | 
| 322 |     // X,Y,T are two parallel view matrices that accumulate two bounding boxes as they map points: | 
| 323 |     // device-space bounds and "45 degree" device-space bounds (| 1 -1 | * devCoords). | 
| 324 |     //                                                          | 1  1 | | 
| 325 |     Sk4f X = Sk4f(m.getScaleX(), m.getSkewY(), m45.getScaleX(), m45.getSkewY()); | 
| 326 |     Sk4f Y = Sk4f(m.getSkewX(), m.getScaleY(), m45.getSkewX(), m45.getScaleY()); | 
| 327 |     Sk4f T = Sk4f(m.getTranslateX(), m.getTranslateY(), m45.getTranslateX(), m45.getTranslateY()); | 
| 328 |  | 
| 329 |     // Map the path's points to device space and accumulate bounding boxes. | 
| 330 |     Sk4f devPt = SkNx_fma(Y, Sk4f(pts[0].y()), T); | 
| 331 |     devPt = SkNx_fma(X, Sk4f(pts[0].x()), devPt); | 
| 332 |     Sk4f topLeft = devPt; | 
| 333 |     Sk4f bottomRight = devPt; | 
| 334 |  | 
| 335 |     // Store all 4 values [dev.x, dev.y, dev45.x, dev45.y]. We are only interested in the first two, | 
| 336 |     // and will overwrite [dev45.x, dev45.y] with the next point. This is why the dst buffer must | 
| 337 |     // be at least one larger than the number of points. | 
| 338 |     devPt.store(&outDevPts[0]); | 
| 339 |  | 
| 340 |     for (int i = 1; i < numPts; ++i) { | 
| 341 |         devPt = SkNx_fma(Y, Sk4f(pts[i].y()), T); | 
| 342 |         devPt = SkNx_fma(X, Sk4f(pts[i].x()), devPt); | 
| 343 |         topLeft = Sk4f::Min(topLeft, devPt); | 
| 344 |         bottomRight = Sk4f::Max(bottomRight, devPt); | 
| 345 |         devPt.store(&outDevPts[i]); | 
| 346 |     } | 
| 347 |  | 
| 348 |     if (!(Sk4f(0) == topLeft*0).allTrue() || !(Sk4f(0) == bottomRight*0).allTrue()) { | 
| 349 |         // The bounds are infinite or NaN. | 
| 350 |         return false; | 
| 351 |     } | 
| 352 |  | 
| 353 |     SkPoint topLeftPts[2], bottomRightPts[2]; | 
| 354 |     topLeft.store(topLeftPts); | 
| 355 |     bottomRight.store(bottomRightPts); | 
| 356 |  | 
| 357 |     const SkRect& devBounds = SkRect::MakeLTRB( | 
| 358 |             topLeftPts[0].x(), topLeftPts[0].y(), bottomRightPts[0].x(), bottomRightPts[0].y()); | 
| 359 |     const SkRect& devBounds45 = SkRect::MakeLTRB( | 
| 360 |             topLeftPts[1].x(), topLeftPts[1].y(), bottomRightPts[1].x(), bottomRightPts[1].y()); | 
| 361 |  | 
| 362 |     octoBounds->set(devBounds, devBounds45); | 
| 363 |     return true; | 
| 364 | } | 
| 365 |  | 
| 366 | GrCCAtlas* GrCCPerFlushResources::renderShapeInAtlas( | 
| 367 |         const SkIRect& clipIBounds, const SkMatrix& m, const GrStyledShape& shape, | 
| 368 |         float strokeDevWidth, GrOctoBounds* octoBounds, SkIRect* devIBounds, | 
| 369 |         SkIVector* devToAtlasOffset) { | 
| 370 |     SkASSERT(this->isMapped()); | 
| 371 |     SkASSERT(fNextPathInstanceIdx < fEndPathInstance); | 
| 372 |  | 
| 373 |     SkPath path; | 
| 374 |     shape.asPath(&path); | 
| 375 |     if (path.isEmpty()) { | 
| 376 |         SkDEBUGCODE(--fEndPathInstance); | 
| 377 |         SkDEBUGCODE(--fEndStencilResolveInstance); | 
| 378 |         return nullptr; | 
| 379 |     } | 
| 380 |     if (!transform_path_pts(m, path, fLocalDevPtsBuffer, octoBounds)) { | 
| 381 |         // The transformed path had infinite or NaN bounds. | 
| 382 |         SkDEBUGCODE(--fEndPathInstance); | 
| 383 |         SkDEBUGCODE(--fEndStencilResolveInstance); | 
| 384 |         return nullptr; | 
| 385 |     } | 
| 386 |  | 
| 387 |     const SkStrokeRec& stroke = shape.style().strokeRec(); | 
| 388 |     if (!stroke.isFillStyle()) { | 
| 389 |         float r = SkStrokeRec::GetInflationRadius( | 
| 390 |                 stroke.getJoin(), stroke.getMiter(), stroke.getCap(), strokeDevWidth); | 
| 391 |         octoBounds->outset(r); | 
| 392 |     } | 
| 393 |  | 
| 394 |     GrScissorTest enableScissorInAtlas; | 
| 395 |     if (clipIBounds.contains(octoBounds->bounds())) { | 
| 396 |         enableScissorInAtlas = GrScissorTest::kDisabled; | 
| 397 |     } else if (octoBounds->clip(clipIBounds)) { | 
| 398 |         enableScissorInAtlas = GrScissorTest::kEnabled; | 
| 399 |     } else { | 
| 400 |         // The clip and octo bounds do not intersect. Draw nothing. | 
| 401 |         SkDEBUGCODE(--fEndPathInstance); | 
| 402 |         SkDEBUGCODE(--fEndStencilResolveInstance); | 
| 403 |         return nullptr; | 
| 404 |     } | 
| 405 |     octoBounds->roundOut(devIBounds); | 
| 406 |     SkASSERT(clipIBounds.contains(*devIBounds)); | 
| 407 |  | 
| 408 |     this->placeRenderedPathInAtlas(*devIBounds, enableScissorInAtlas, devToAtlasOffset); | 
| 409 |  | 
| 410 |     GrFillRule fillRule; | 
| 411 |     if (stroke.isFillStyle()) { | 
| 412 |         SkASSERT(0 == strokeDevWidth); | 
| 413 |         fFiller.parseDeviceSpaceFill(path, fLocalDevPtsBuffer.begin(), enableScissorInAtlas, | 
| 414 |                                      *devIBounds, *devToAtlasOffset); | 
| 415 |         fillRule = GrFillRuleForSkPath(path); | 
| 416 |     } else { | 
| 417 |         // Stroke-and-fill is not yet supported. | 
| 418 |         SkASSERT(SkStrokeRec::kStroke_Style == stroke.getStyle() || stroke.isHairlineStyle()); | 
| 419 |         SkASSERT(!stroke.isHairlineStyle() || 1 == strokeDevWidth); | 
| 420 |         fStroker.parseDeviceSpaceStroke( | 
| 421 |                 path, fLocalDevPtsBuffer.begin(), stroke, strokeDevWidth, enableScissorInAtlas, | 
| 422 |                 *devIBounds, *devToAtlasOffset); | 
| 423 |         fillRule = GrFillRule::kNonzero; | 
| 424 |     } | 
| 425 |  | 
| 426 |     if (GrCCAtlas::CoverageType::kA8_Multisample == this->renderedPathCoverageType()) { | 
| 427 |         this->recordStencilResolveInstance(*devIBounds, *devToAtlasOffset, fillRule); | 
| 428 |     } | 
| 429 |  | 
| 430 |     return &fRenderedAtlasStack.current(); | 
| 431 | } | 
| 432 |  | 
| 433 | const GrCCAtlas* GrCCPerFlushResources::renderDeviceSpacePathInAtlas( | 
| 434 |         const SkIRect& clipIBounds, const SkPath& devPath, const SkIRect& devPathIBounds, | 
| 435 |         GrFillRule fillRule, SkIVector* devToAtlasOffset) { | 
| 436 |     SkASSERT(this->isMapped()); | 
| 437 |  | 
| 438 |     if (devPath.isEmpty()) { | 
| 439 |         SkDEBUGCODE(--fEndStencilResolveInstance); | 
| 440 |         return nullptr; | 
| 441 |     } | 
| 442 |  | 
| 443 |     GrScissorTest enableScissorInAtlas; | 
| 444 |     SkIRect clippedPathIBounds; | 
| 445 |     if (clipIBounds.contains(devPathIBounds)) { | 
| 446 |         clippedPathIBounds = devPathIBounds; | 
| 447 |         enableScissorInAtlas = GrScissorTest::kDisabled; | 
| 448 |     } else if (clippedPathIBounds.intersect(clipIBounds, devPathIBounds)) { | 
| 449 |         enableScissorInAtlas = GrScissorTest::kEnabled; | 
| 450 |     } else { | 
| 451 |         // The clip and path bounds do not intersect. Draw nothing. | 
| 452 |         SkDEBUGCODE(--fEndStencilResolveInstance); | 
| 453 |         return nullptr; | 
| 454 |     } | 
| 455 |  | 
| 456 |     this->placeRenderedPathInAtlas(clippedPathIBounds, enableScissorInAtlas, devToAtlasOffset); | 
| 457 |     fFiller.parseDeviceSpaceFill(devPath, SkPathPriv::PointData(devPath), enableScissorInAtlas, | 
| 458 |                                  clippedPathIBounds, *devToAtlasOffset); | 
| 459 |  | 
| 460 |     // In MSAA mode we also record an internal draw instance that will be used to resolve stencil | 
| 461 |     // winding values to coverage when the atlas is generated. | 
| 462 |     if (GrCCAtlas::CoverageType::kA8_Multisample == this->renderedPathCoverageType()) { | 
| 463 |         this->recordStencilResolveInstance(clippedPathIBounds, *devToAtlasOffset, fillRule); | 
| 464 |     } | 
| 465 |  | 
| 466 |     return &fRenderedAtlasStack.current(); | 
| 467 | } | 
| 468 |  | 
| 469 | void GrCCPerFlushResources::placeRenderedPathInAtlas( | 
| 470 |         const SkIRect& clippedPathIBounds, GrScissorTest scissorTest, SkIVector* devToAtlasOffset) { | 
| 471 |     if (GrCCAtlas* retiredAtlas = | 
| 472 |                 fRenderedAtlasStack.addRect(clippedPathIBounds, devToAtlasOffset)) { | 
| 473 |         // We did not fit in the previous coverage count atlas and it was retired. Close the path | 
| 474 |         // parser's current batch (which does not yet include the path we just parsed). We will | 
| 475 |         // render this batch into the retired atlas during finalize(). | 
| 476 |         retiredAtlas->setFillBatchID(fFiller.closeCurrentBatch()); | 
| 477 |         retiredAtlas->setStrokeBatchID(fStroker.closeCurrentBatch()); | 
| 478 |         retiredAtlas->setEndStencilResolveInstance(fNextStencilResolveInstanceIdx); | 
| 479 |     } | 
| 480 | } | 
| 481 |  | 
| 482 | void GrCCPerFlushResources::recordStencilResolveInstance( | 
| 483 |         const SkIRect& clippedPathIBounds, const SkIVector& devToAtlasOffset, GrFillRule fillRule) { | 
| 484 |     SkASSERT(GrCCAtlas::CoverageType::kA8_Multisample == this->renderedPathCoverageType()); | 
| 485 |     SkASSERT(fNextStencilResolveInstanceIdx < fEndStencilResolveInstance); | 
| 486 |  | 
| 487 |     SkIRect atlasIBounds = clippedPathIBounds.makeOffset(devToAtlasOffset); | 
| 488 |     if (GrFillRule::kEvenOdd == fillRule) { | 
| 489 |         // Make even/odd fills counterclockwise. The resolve draw uses two-sided stencil, with | 
| 490 |         // "nonzero" settings in front and "even/odd" settings in back. | 
| 491 |         std::swap(atlasIBounds.fLeft, atlasIBounds.fRight); | 
| 492 |     } | 
| 493 |     fStencilResolveBuffer[fNextStencilResolveInstanceIdx++] = { | 
| 494 |             (int16_t)atlasIBounds.left(), (int16_t)atlasIBounds.top(), | 
| 495 |             (int16_t)atlasIBounds.right(), (int16_t)atlasIBounds.bottom()}; | 
| 496 | } | 
| 497 |  | 
| 498 | bool GrCCPerFlushResources::finalize(GrOnFlushResourceProvider* onFlushRP) { | 
| 499 |     SkASSERT(this->isMapped()); | 
| 500 |     SkASSERT(fNextPathInstanceIdx == fEndPathInstance); | 
| 501 |     SkASSERT(fNextCopyInstanceIdx == fEndCopyInstance); | 
| 502 |     SkASSERT(GrCCAtlas::CoverageType::kA8_Multisample != this->renderedPathCoverageType() || | 
| 503 |              fNextStencilResolveInstanceIdx == fEndStencilResolveInstance); | 
| 504 |  | 
| 505 |     fPathInstanceBuffer.unmapBuffer(); | 
| 506 |  | 
| 507 |     if (fStencilResolveBuffer.hasGpuBuffer()) { | 
| 508 |         fStencilResolveBuffer.unmapBuffer(); | 
| 509 |     } | 
| 510 |  | 
| 511 |     if (!fCopyAtlasStack.empty()) { | 
| 512 |         fCopyAtlasStack.current().setFillBatchID(fCopyPathRanges.count()); | 
| 513 |         fCurrCopyAtlasRangesIdx = fCopyPathRanges.count(); | 
| 514 |     } | 
| 515 |     if (!fRenderedAtlasStack.empty()) { | 
| 516 |         fRenderedAtlasStack.current().setFillBatchID(fFiller.closeCurrentBatch()); | 
| 517 |         fRenderedAtlasStack.current().setStrokeBatchID(fStroker.closeCurrentBatch()); | 
| 518 |         fRenderedAtlasStack.current().setEndStencilResolveInstance(fNextStencilResolveInstanceIdx); | 
| 519 |     } | 
| 520 |  | 
| 521 |     // Build the GPU buffers to render path coverage counts. (This must not happen until after the | 
| 522 |     // final calls to fFiller/fStroker.closeCurrentBatch().) | 
| 523 |     if (!fFiller.prepareToDraw(onFlushRP)) { | 
| 524 |         return false; | 
| 525 |     } | 
| 526 |     if (!fStroker.prepareToDraw(onFlushRP)) { | 
| 527 |         return false; | 
| 528 |     } | 
| 529 |  | 
| 530 |     // Draw the copies from coverage count or msaa atlas(es) into 8-bit cached atlas(es). | 
| 531 |     int copyRangeIdx = 0; | 
| 532 |     int baseCopyInstance = 0; | 
| 533 |     for (GrCCAtlas& atlas : fCopyAtlasStack.atlases()) { | 
| 534 |         int endCopyRange = atlas.getFillBatchID(); | 
| 535 |         SkASSERT(endCopyRange > copyRangeIdx); | 
| 536 |  | 
| 537 |         auto rtc = atlas.instantiate(onFlushRP); | 
| 538 |         for (; copyRangeIdx < endCopyRange; ++copyRangeIdx) { | 
| 539 |             const CopyPathRange& copyRange = fCopyPathRanges[copyRangeIdx]; | 
| 540 |             int endCopyInstance = baseCopyInstance + copyRange.fCount; | 
| 541 |             if (rtc) { | 
| 542 |                 auto op = CopyAtlasOp::Make( | 
| 543 |                         rtc->surfPriv().getContext(), sk_ref_sp(this), copyRange.fSrcProxy, | 
| 544 |                         baseCopyInstance, endCopyInstance, atlas.drawBounds()); | 
| 545 |                 rtc->addDrawOp(nullptr, std::move(op)); | 
| 546 |             } | 
| 547 |             baseCopyInstance = endCopyInstance; | 
| 548 |         } | 
| 549 |     } | 
| 550 |     SkASSERT(fCopyPathRanges.count() == copyRangeIdx); | 
| 551 |     SkASSERT(fNextCopyInstanceIdx == baseCopyInstance); | 
| 552 |     SkASSERT(baseCopyInstance == fEndCopyInstance); | 
| 553 |  | 
| 554 |     // Render the coverage count atlas(es). | 
| 555 |     int baseStencilResolveInstance = 0; | 
| 556 |     for (GrCCAtlas& atlas : fRenderedAtlasStack.atlases()) { | 
| 557 |         // Copies will be finished by the time we get to rendering new atlases. See if we can | 
| 558 |         // recycle any previous invalidated atlas textures instead of creating new ones. | 
| 559 |         sk_sp<GrTexture> backingTexture; | 
| 560 |         for (sk_sp<GrTexture>& texture : fRecyclableAtlasTextures) { | 
| 561 |             if (texture && atlas.currentHeight() == texture->height() && | 
| 562 |                     atlas.currentWidth() == texture->width()) { | 
| 563 |                 backingTexture = std::exchange(texture, nullptr); | 
| 564 |                 break; | 
| 565 |             } | 
| 566 |         } | 
| 567 |  | 
| 568 |         if (auto rtc = atlas.instantiate(onFlushRP, std::move(backingTexture))) { | 
| 569 |             std::unique_ptr<GrDrawOp> op; | 
| 570 |             if (CoverageType::kA8_Multisample == fRenderedAtlasStack.coverageType()) { | 
| 571 |                 op = GrStencilAtlasOp::Make( | 
| 572 |                         rtc->surfPriv().getContext(), sk_ref_sp(this), atlas.getFillBatchID(), | 
| 573 |                         atlas.getStrokeBatchID(), baseStencilResolveInstance, | 
| 574 |                         atlas.getEndStencilResolveInstance(), atlas.drawBounds()); | 
| 575 |             } else if (onFlushRP->caps()->shaderCaps()->geometryShaderSupport()) { | 
| 576 |                 op = RenderAtlasOp<GrGSCoverageProcessor>::Make( | 
| 577 |                         rtc->surfPriv().getContext(), sk_ref_sp(this), atlas.getFillBatchID(), | 
| 578 |                         atlas.getStrokeBatchID(), atlas.drawBounds()); | 
| 579 |             } else { | 
| 580 |                 op = RenderAtlasOp<GrVSCoverageProcessor>::Make( | 
| 581 |                         rtc->surfPriv().getContext(), sk_ref_sp(this), atlas.getFillBatchID(), | 
| 582 |                         atlas.getStrokeBatchID(), atlas.drawBounds()); | 
| 583 |             } | 
| 584 |             rtc->addDrawOp(nullptr, std::move(op)); | 
| 585 |             if (rtc->asSurfaceProxy()->requiresManualMSAAResolve()) { | 
| 586 |                 onFlushRP->addTextureResolveTask(sk_ref_sp(rtc->asTextureProxy()), | 
| 587 |                                                  GrSurfaceProxy::ResolveFlags::kMSAA); | 
| 588 |             } | 
| 589 |         } | 
| 590 |  | 
| 591 |         SkASSERT(atlas.getEndStencilResolveInstance() >= baseStencilResolveInstance); | 
| 592 |         baseStencilResolveInstance = atlas.getEndStencilResolveInstance(); | 
| 593 |     } | 
| 594 |     SkASSERT(GrCCAtlas::CoverageType::kA8_Multisample != this->renderedPathCoverageType() || | 
| 595 |              baseStencilResolveInstance == fEndStencilResolveInstance); | 
| 596 |  | 
| 597 |     return true; | 
| 598 | } | 
| 599 |  | 
| 600 | void GrCCPerFlushResourceSpecs::cancelCopies() { | 
| 601 |     // Convert copies to cached draws. | 
| 602 |     fNumCachedPaths += fNumCopiedPaths[kFillIdx] + fNumCopiedPaths[kStrokeIdx]; | 
| 603 |     fNumCopiedPaths[kFillIdx] = fNumCopiedPaths[kStrokeIdx] = 0; | 
| 604 |     fCopyPathStats[kFillIdx] = fCopyPathStats[kStrokeIdx] = GrCCRenderedPathStats(); | 
| 605 |     fCopyAtlasSpecs = GrCCAtlas::Specs(); | 
| 606 | } | 
| 607 |  |