| 1 | /* | 
|---|
| 2 | * Copyright © 2020  Google, Inc. | 
|---|
| 3 | * | 
|---|
| 4 | *  This is part of HarfBuzz, a text shaping library. | 
|---|
| 5 | * | 
|---|
| 6 | * Permission is hereby granted, without written agreement and without | 
|---|
| 7 | * license or royalty fees, to use, copy, modify, and distribute this | 
|---|
| 8 | * software and its documentation for any purpose, provided that the | 
|---|
| 9 | * above copyright notice and the following two paragraphs appear in | 
|---|
| 10 | * all copies of this software. | 
|---|
| 11 | * | 
|---|
| 12 | * IN NO EVENT SHALL THE COPYRIGHT HOLDER BE LIABLE TO ANY PARTY FOR | 
|---|
| 13 | * DIRECT, INDIRECT, SPECIAL, INCIDENTAL, OR CONSEQUENTIAL DAMAGES | 
|---|
| 14 | * ARISING OUT OF THE USE OF THIS SOFTWARE AND ITS DOCUMENTATION, EVEN | 
|---|
| 15 | * IF THE COPYRIGHT HOLDER HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH | 
|---|
| 16 | * DAMAGE. | 
|---|
| 17 | * | 
|---|
| 18 | * THE COPYRIGHT HOLDER SPECIFICALLY DISCLAIMS ANY WARRANTIES, INCLUDING, | 
|---|
| 19 | * BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND | 
|---|
| 20 | * FITNESS FOR A PARTICULAR PURPOSE.  THE SOFTWARE PROVIDED HEREUNDER IS | 
|---|
| 21 | * ON AN "AS IS" BASIS, AND THE COPYRIGHT HOLDER HAS NO OBLIGATION TO | 
|---|
| 22 | * PROVIDE MAINTENANCE, SUPPORT, UPDATES, ENHANCEMENTS, OR MODIFICATIONS. | 
|---|
| 23 | * | 
|---|
| 24 | * Google Author(s): Garret Rieger | 
|---|
| 25 | */ | 
|---|
| 26 |  | 
|---|
| 27 | #ifndef HB_REPACKER_HH | 
|---|
| 28 | #define HB_REPACKER_HH | 
|---|
| 29 |  | 
|---|
| 30 | #include "hb-open-type.hh" | 
|---|
| 31 | #include "hb-map.hh" | 
|---|
| 32 | #include "hb-vector.hh" | 
|---|
| 33 | #include "graph/graph.hh" | 
|---|
| 34 | #include "graph/gsubgpos-graph.hh" | 
|---|
| 35 | #include "graph/serialize.hh" | 
|---|
| 36 |  | 
|---|
| 37 | using graph::graph_t; | 
|---|
| 38 |  | 
|---|
| 39 | /* | 
|---|
| 40 | * For a detailed writeup on the overflow resolution algorithm see: | 
|---|
| 41 | * docs/repacker.md | 
|---|
| 42 | */ | 
|---|
| 43 |  | 
|---|
| 44 | struct lookup_size_t | 
|---|
| 45 | { | 
|---|
| 46 | unsigned lookup_index; | 
|---|
| 47 | size_t size; | 
|---|
| 48 | unsigned num_subtables; | 
|---|
| 49 |  | 
|---|
| 50 | static int cmp (const void* a, const void* b) | 
|---|
| 51 | { | 
|---|
| 52 | return cmp ((const lookup_size_t*) a, | 
|---|
| 53 | (const lookup_size_t*) b); | 
|---|
| 54 | } | 
|---|
| 55 |  | 
|---|
| 56 | static int cmp (const lookup_size_t* a, const lookup_size_t* b) | 
|---|
| 57 | { | 
|---|
| 58 | double subtables_per_byte_a = (double) a->num_subtables / (double) a->size; | 
|---|
| 59 | double subtables_per_byte_b = (double) b->num_subtables / (double) b->size; | 
|---|
| 60 | if (subtables_per_byte_a == subtables_per_byte_b) { | 
|---|
| 61 | return b->lookup_index - a->lookup_index; | 
|---|
| 62 | } | 
|---|
| 63 |  | 
|---|
| 64 | double cmp = subtables_per_byte_b - subtables_per_byte_a; | 
|---|
| 65 | if (cmp < 0) return -1; | 
|---|
| 66 | if (cmp > 0) return 1; | 
|---|
| 67 | return 0; | 
|---|
| 68 | } | 
|---|
| 69 | }; | 
|---|
| 70 |  | 
|---|
| 71 | static inline | 
|---|
| 72 | bool _presplit_subtables_if_needed (graph::gsubgpos_graph_context_t& ext_context) | 
|---|
| 73 | { | 
|---|
| 74 | // For each lookup this will check the size of subtables and split them as needed | 
|---|
| 75 | // so that no subtable is at risk of overflowing. (where we support splitting for | 
|---|
| 76 | // that subtable type). | 
|---|
| 77 | // | 
|---|
| 78 | // TODO(grieger): de-dup newly added nodes as necessary. Probably just want a full de-dup | 
|---|
| 79 | //                pass after this processing is done. Not super necessary as splits are | 
|---|
| 80 | //                only done where overflow is likely, so de-dup probably will get undone | 
|---|
| 81 | //                later anyways. | 
|---|
| 82 | for (unsigned lookup_index : ext_context.lookups.keys ()) | 
|---|
| 83 | { | 
|---|
| 84 | graph::Lookup* lookup = ext_context.lookups.get(lookup_index); | 
|---|
| 85 | if (!lookup->split_subtables_if_needed (ext_context, lookup_index)) | 
|---|
| 86 | return false; | 
|---|
| 87 | } | 
|---|
| 88 |  | 
|---|
| 89 | return true; | 
|---|
| 90 | } | 
|---|
| 91 |  | 
|---|
| 92 | /* | 
|---|
| 93 | * Analyze the lookups in a GSUB/GPOS table and decide if any should be promoted | 
|---|
| 94 | * to extension lookups. | 
|---|
| 95 | */ | 
|---|
| 96 | static inline | 
|---|
| 97 | bool _promote_extensions_if_needed (graph::gsubgpos_graph_context_t& ext_context) | 
|---|
| 98 | { | 
|---|
| 99 | // Simple Algorithm (v1, current): | 
|---|
| 100 | // 1. Calculate how many bytes each non-extension lookup consumes. | 
|---|
| 101 | // 2. Select up to 64k of those to remain as non-extension (greedy, highest subtables per byte first) | 
|---|
| 102 | // 3. Promote the rest. | 
|---|
| 103 | // | 
|---|
| 104 | // Advanced Algorithm (v2, not implemented): | 
|---|
| 105 | // 1. Perform connected component analysis using lookups as roots. | 
|---|
| 106 | // 2. Compute size of each connected component. | 
|---|
| 107 | // 3. Select up to 64k worth of connected components to remain as non-extensions. | 
|---|
| 108 | //    (greedy, highest subtables per byte first) | 
|---|
| 109 | // 4. Promote the rest. | 
|---|
| 110 |  | 
|---|
| 111 | // TODO(garretrieger): support extension demotion, then consider all lookups. Requires advanced algo. | 
|---|
| 112 | // TODO(garretrieger): also support extension promotion during iterative resolution phase, then | 
|---|
| 113 | //                     we can use a less conservative threshold here. | 
|---|
| 114 | // TODO(grieger): skip this for the 24 bit case. | 
|---|
| 115 | if (!ext_context.lookups) return true; | 
|---|
| 116 |  | 
|---|
| 117 | hb_vector_t<lookup_size_t> lookup_sizes; | 
|---|
| 118 | lookup_sizes.alloc (ext_context.lookups.get_population (), true); | 
|---|
| 119 |  | 
|---|
| 120 | for (unsigned lookup_index : ext_context.lookups.keys ()) | 
|---|
| 121 | { | 
|---|
| 122 | const graph::Lookup* lookup = ext_context.lookups.get(lookup_index); | 
|---|
| 123 | hb_set_t visited; | 
|---|
| 124 | lookup_sizes.push (lookup_size_t { | 
|---|
| 125 | lookup_index, | 
|---|
| 126 | ext_context.graph.find_subgraph_size (lookup_index, visited), | 
|---|
| 127 | lookup->number_of_subtables (), | 
|---|
| 128 | }); | 
|---|
| 129 | } | 
|---|
| 130 |  | 
|---|
| 131 | lookup_sizes.qsort (); | 
|---|
| 132 |  | 
|---|
| 133 | size_t lookup_list_size = ext_context.graph.vertices_[ext_context.lookup_list_index].table_size (); | 
|---|
| 134 | size_t l2_l3_size = lookup_list_size; // Lookup List + Lookups | 
|---|
| 135 | size_t l3_l4_size = 0; // Lookups + SubTables | 
|---|
| 136 | size_t l4_plus_size = 0; // SubTables + their descendants | 
|---|
| 137 |  | 
|---|
| 138 | // Start by assuming all lookups are using extension subtables, this size will be removed later | 
|---|
| 139 | // if it's decided to not make a lookup extension. | 
|---|
| 140 | for (auto p : lookup_sizes) | 
|---|
| 141 | { | 
|---|
| 142 | unsigned subtables_size = p.num_subtables * 8; | 
|---|
| 143 | l3_l4_size += subtables_size; | 
|---|
| 144 | l4_plus_size += subtables_size; | 
|---|
| 145 | } | 
|---|
| 146 |  | 
|---|
| 147 | bool layers_full = false; | 
|---|
| 148 | for (auto p : lookup_sizes) | 
|---|
| 149 | { | 
|---|
| 150 | const graph::Lookup* lookup = ext_context.lookups.get(p.lookup_index); | 
|---|
| 151 | if (lookup->is_extension (ext_context.table_tag)) | 
|---|
| 152 | // already an extension so size is counted by the loop above. | 
|---|
| 153 | continue; | 
|---|
| 154 |  | 
|---|
| 155 | if (!layers_full) | 
|---|
| 156 | { | 
|---|
| 157 | size_t lookup_size = ext_context.graph.vertices_[p.lookup_index].table_size (); | 
|---|
| 158 | hb_set_t visited; | 
|---|
| 159 | size_t subtables_size = ext_context.graph.find_subgraph_size (p.lookup_index, visited, 1) - lookup_size; | 
|---|
| 160 | size_t remaining_size = p.size - subtables_size - lookup_size; | 
|---|
| 161 |  | 
|---|
| 162 | l2_l3_size   += lookup_size; | 
|---|
| 163 | l3_l4_size   += lookup_size + subtables_size; | 
|---|
| 164 | l3_l4_size   -= p.num_subtables * 8; | 
|---|
| 165 | l4_plus_size += subtables_size + remaining_size; | 
|---|
| 166 |  | 
|---|
| 167 | if (l2_l3_size < (1 << 16) | 
|---|
| 168 | && l3_l4_size < (1 << 16) | 
|---|
| 169 | && l4_plus_size < (1 << 16)) continue; // this lookup fits within all layers groups | 
|---|
| 170 |  | 
|---|
| 171 | layers_full = true; | 
|---|
| 172 | } | 
|---|
| 173 |  | 
|---|
| 174 | if (!ext_context.lookups.get(p.lookup_index)->make_extension (ext_context, p.lookup_index)) | 
|---|
| 175 | return false; | 
|---|
| 176 | } | 
|---|
| 177 |  | 
|---|
| 178 | return true; | 
|---|
| 179 | } | 
|---|
| 180 |  | 
|---|
| 181 | static inline | 
|---|
| 182 | bool _try_isolating_subgraphs (const hb_vector_t<graph::overflow_record_t>& overflows, | 
|---|
| 183 | graph_t& sorted_graph) | 
|---|
| 184 | { | 
|---|
| 185 | unsigned space = 0; | 
|---|
| 186 | hb_set_t roots_to_isolate; | 
|---|
| 187 |  | 
|---|
| 188 | for (int i = overflows.length - 1; i >= 0; i--) | 
|---|
| 189 | { | 
|---|
| 190 | const graph::overflow_record_t& r = overflows[i]; | 
|---|
| 191 |  | 
|---|
| 192 | unsigned root; | 
|---|
| 193 | unsigned overflow_space = sorted_graph.space_for (r.parent, &root); | 
|---|
| 194 | if (!overflow_space) continue; | 
|---|
| 195 | if (sorted_graph.num_roots_for_space (overflow_space) <= 1) continue; | 
|---|
| 196 |  | 
|---|
| 197 | if (!space) { | 
|---|
| 198 | space = overflow_space; | 
|---|
| 199 | } | 
|---|
| 200 |  | 
|---|
| 201 | if (space == overflow_space) | 
|---|
| 202 | roots_to_isolate.add(root); | 
|---|
| 203 | } | 
|---|
| 204 |  | 
|---|
| 205 | if (!roots_to_isolate) return false; | 
|---|
| 206 |  | 
|---|
| 207 | unsigned maximum_to_move = hb_max ((sorted_graph.num_roots_for_space (space) / 2u), 1u); | 
|---|
| 208 | if (roots_to_isolate.get_population () > maximum_to_move) { | 
|---|
| 209 | // Only move at most half of the roots in a space at a time. | 
|---|
| 210 | unsigned  = roots_to_isolate.get_population () - maximum_to_move; | 
|---|
| 211 | while (extra--) { | 
|---|
| 212 | uint32_t root = HB_SET_VALUE_INVALID; | 
|---|
| 213 | roots_to_isolate.previous (&root); | 
|---|
| 214 | roots_to_isolate.del (root); | 
|---|
| 215 | } | 
|---|
| 216 | } | 
|---|
| 217 |  | 
|---|
| 218 | DEBUG_MSG (SUBSET_REPACK, nullptr, | 
|---|
| 219 | "Overflow in space %u (%u roots). Moving %u roots to space %u.", | 
|---|
| 220 | space, | 
|---|
| 221 | sorted_graph.num_roots_for_space (space), | 
|---|
| 222 | roots_to_isolate.get_population (), | 
|---|
| 223 | sorted_graph.next_space ()); | 
|---|
| 224 |  | 
|---|
| 225 | sorted_graph.isolate_subgraph (roots_to_isolate); | 
|---|
| 226 | sorted_graph.move_to_new_space (roots_to_isolate); | 
|---|
| 227 |  | 
|---|
| 228 | return true; | 
|---|
| 229 | } | 
|---|
| 230 |  | 
|---|
| 231 | static inline | 
|---|
| 232 | bool _process_overflows (const hb_vector_t<graph::overflow_record_t>& overflows, | 
|---|
| 233 | hb_set_t& priority_bumped_parents, | 
|---|
| 234 | graph_t& sorted_graph) | 
|---|
| 235 | { | 
|---|
| 236 | bool resolution_attempted = false; | 
|---|
| 237 |  | 
|---|
| 238 | // Try resolving the furthest overflows first. | 
|---|
| 239 | for (int i = overflows.length - 1; i >= 0; i--) | 
|---|
| 240 | { | 
|---|
| 241 | const graph::overflow_record_t& r = overflows[i]; | 
|---|
| 242 | const auto& child = sorted_graph.vertices_[r.child]; | 
|---|
| 243 | if (child.is_shared ()) | 
|---|
| 244 | { | 
|---|
| 245 | // The child object is shared, we may be able to eliminate the overflow | 
|---|
| 246 | // by duplicating it. | 
|---|
| 247 | if (sorted_graph.duplicate (r.parent, r.child) == (unsigned) -1) continue; | 
|---|
| 248 | return true; | 
|---|
| 249 | } | 
|---|
| 250 |  | 
|---|
| 251 | if (child.is_leaf () && !priority_bumped_parents.has (r.parent)) | 
|---|
| 252 | { | 
|---|
| 253 | // This object is too far from it's parent, attempt to move it closer. | 
|---|
| 254 | // | 
|---|
| 255 | // TODO(garretrieger): initially limiting this to leaf's since they can be | 
|---|
| 256 | //                     moved closer with fewer consequences. However, this can | 
|---|
| 257 | //                     likely can be used for non-leafs as well. | 
|---|
| 258 | // TODO(garretrieger): also try lowering priority of the parent. Make it | 
|---|
| 259 | //                     get placed further up in the ordering, closer to it's children. | 
|---|
| 260 | //                     this is probably preferable if the total size of the parent object | 
|---|
| 261 | //                     is < then the total size of the children (and the parent can be moved). | 
|---|
| 262 | //                     Since in that case moving the parent will cause a smaller increase in | 
|---|
| 263 | //                     the length of other offsets. | 
|---|
| 264 | if (sorted_graph.raise_childrens_priority (r.parent)) { | 
|---|
| 265 | priority_bumped_parents.add (r.parent); | 
|---|
| 266 | resolution_attempted = true; | 
|---|
| 267 | } | 
|---|
| 268 | continue; | 
|---|
| 269 | } | 
|---|
| 270 |  | 
|---|
| 271 | // TODO(garretrieger): add additional offset resolution strategies | 
|---|
| 272 | // - Promotion to extension lookups. | 
|---|
| 273 | // - Table splitting. | 
|---|
| 274 | } | 
|---|
| 275 |  | 
|---|
| 276 | return resolution_attempted; | 
|---|
| 277 | } | 
|---|
| 278 |  | 
|---|
| 279 | inline bool | 
|---|
| 280 | hb_resolve_graph_overflows (hb_tag_t table_tag, | 
|---|
| 281 | unsigned max_rounds , | 
|---|
| 282 | bool recalculate_extensions, | 
|---|
| 283 | graph_t& sorted_graph /* IN/OUT */) | 
|---|
| 284 | { | 
|---|
| 285 | sorted_graph.sort_shortest_distance (); | 
|---|
| 286 | if (sorted_graph.in_error ()) | 
|---|
| 287 | { | 
|---|
| 288 | DEBUG_MSG (SUBSET_REPACK, nullptr, "Sorted graph in error state after initial sort."); | 
|---|
| 289 | return false; | 
|---|
| 290 | } | 
|---|
| 291 |  | 
|---|
| 292 | bool will_overflow = graph::will_overflow (sorted_graph); | 
|---|
| 293 | if (!will_overflow) | 
|---|
| 294 | return true; | 
|---|
| 295 |  | 
|---|
| 296 | graph::gsubgpos_graph_context_t ext_context (table_tag, sorted_graph); | 
|---|
| 297 | if ((table_tag == HB_OT_TAG_GPOS | 
|---|
| 298 | ||  table_tag == HB_OT_TAG_GSUB) | 
|---|
| 299 | && will_overflow) | 
|---|
| 300 | { | 
|---|
| 301 | if (recalculate_extensions) | 
|---|
| 302 | { | 
|---|
| 303 | DEBUG_MSG (SUBSET_REPACK, nullptr, "Splitting subtables if needed."); | 
|---|
| 304 | if (!_presplit_subtables_if_needed (ext_context)) { | 
|---|
| 305 | DEBUG_MSG (SUBSET_REPACK, nullptr, "Subtable splitting failed."); | 
|---|
| 306 | return false; | 
|---|
| 307 | } | 
|---|
| 308 |  | 
|---|
| 309 | DEBUG_MSG (SUBSET_REPACK, nullptr, "Promoting lookups to extensions if needed."); | 
|---|
| 310 | if (!_promote_extensions_if_needed (ext_context)) { | 
|---|
| 311 | DEBUG_MSG (SUBSET_REPACK, nullptr, "Extensions promotion failed."); | 
|---|
| 312 | return false; | 
|---|
| 313 | } | 
|---|
| 314 | } | 
|---|
| 315 |  | 
|---|
| 316 | DEBUG_MSG (SUBSET_REPACK, nullptr, "Assigning spaces to 32 bit subgraphs."); | 
|---|
| 317 | if (sorted_graph.assign_spaces ()) | 
|---|
| 318 | sorted_graph.sort_shortest_distance (); | 
|---|
| 319 | else | 
|---|
| 320 | sorted_graph.sort_shortest_distance_if_needed (); | 
|---|
| 321 | } | 
|---|
| 322 |  | 
|---|
| 323 | unsigned round = 0; | 
|---|
| 324 | hb_vector_t<graph::overflow_record_t> overflows; | 
|---|
| 325 | // TODO(garretrieger): select a good limit for max rounds. | 
|---|
| 326 | while (!sorted_graph.in_error () | 
|---|
| 327 | && graph::will_overflow (sorted_graph, &overflows) | 
|---|
| 328 | && round < max_rounds) { | 
|---|
| 329 | DEBUG_MSG (SUBSET_REPACK, nullptr, "=== Overflow resolution round %u ===", round); | 
|---|
| 330 | print_overflows (sorted_graph, overflows); | 
|---|
| 331 |  | 
|---|
| 332 | hb_set_t priority_bumped_parents; | 
|---|
| 333 |  | 
|---|
| 334 | if (!_try_isolating_subgraphs (overflows, sorted_graph)) | 
|---|
| 335 | { | 
|---|
| 336 | // Don't count space isolation towards round limit. Only increment | 
|---|
| 337 | // round counter if space isolation made no changes. | 
|---|
| 338 | round++; | 
|---|
| 339 | if (!_process_overflows (overflows, priority_bumped_parents, sorted_graph)) | 
|---|
| 340 | { | 
|---|
| 341 | DEBUG_MSG (SUBSET_REPACK, nullptr, "No resolution available :("); | 
|---|
| 342 | break; | 
|---|
| 343 | } | 
|---|
| 344 | } | 
|---|
| 345 |  | 
|---|
| 346 | sorted_graph.sort_shortest_distance (); | 
|---|
| 347 | } | 
|---|
| 348 |  | 
|---|
| 349 | if (sorted_graph.in_error ()) | 
|---|
| 350 | { | 
|---|
| 351 | DEBUG_MSG (SUBSET_REPACK, nullptr, "Sorted graph in error state."); | 
|---|
| 352 | return false; | 
|---|
| 353 | } | 
|---|
| 354 |  | 
|---|
| 355 | if (graph::will_overflow (sorted_graph)) | 
|---|
| 356 | { | 
|---|
| 357 | DEBUG_MSG (SUBSET_REPACK, nullptr, "Offset overflow resolution failed."); | 
|---|
| 358 | return false; | 
|---|
| 359 | } | 
|---|
| 360 |  | 
|---|
| 361 | return true; | 
|---|
| 362 | } | 
|---|
| 363 |  | 
|---|
| 364 | /* | 
|---|
| 365 | * Attempts to modify the topological sorting of the provided object graph to | 
|---|
| 366 | * eliminate offset overflows in the links between objects of the graph. If a | 
|---|
| 367 | * non-overflowing ordering is found the updated graph is serialized it into the | 
|---|
| 368 | * provided serialization context. | 
|---|
| 369 | * | 
|---|
| 370 | * If necessary the structure of the graph may be modified in ways that do not | 
|---|
| 371 | * affect the functionality of the graph. For example shared objects may be | 
|---|
| 372 | * duplicated. | 
|---|
| 373 | * | 
|---|
| 374 | * For a detailed writeup describing how the algorithm operates see: | 
|---|
| 375 | * docs/repacker.md | 
|---|
| 376 | */ | 
|---|
| 377 | template<typename T> | 
|---|
| 378 | inline hb_blob_t* | 
|---|
| 379 | hb_resolve_overflows (const T& packed, | 
|---|
| 380 | hb_tag_t table_tag, | 
|---|
| 381 | unsigned max_rounds = 20, | 
|---|
| 382 | bool recalculate_extensions = false) { | 
|---|
| 383 | graph_t sorted_graph (packed); | 
|---|
| 384 | if (sorted_graph.in_error ()) | 
|---|
| 385 | { | 
|---|
| 386 | // Invalid graph definition. | 
|---|
| 387 | return nullptr; | 
|---|
| 388 | } | 
|---|
| 389 |  | 
|---|
| 390 | if (!sorted_graph.is_fully_connected ()) | 
|---|
| 391 | { | 
|---|
| 392 | sorted_graph.print_orphaned_nodes (); | 
|---|
| 393 | return nullptr; | 
|---|
| 394 | } | 
|---|
| 395 |  | 
|---|
| 396 | if (sorted_graph.in_error ()) | 
|---|
| 397 | { | 
|---|
| 398 | // Allocations failed somewhere | 
|---|
| 399 | DEBUG_MSG (SUBSET_REPACK, nullptr, | 
|---|
| 400 | "Graph is in error, likely due to a memory allocation error."); | 
|---|
| 401 | return nullptr; | 
|---|
| 402 | } | 
|---|
| 403 |  | 
|---|
| 404 | if (!hb_resolve_graph_overflows (table_tag, max_rounds, recalculate_extensions, sorted_graph)) | 
|---|
| 405 | return nullptr; | 
|---|
| 406 |  | 
|---|
| 407 | return graph::serialize (sorted_graph); | 
|---|
| 408 | } | 
|---|
| 409 |  | 
|---|
| 410 | #endif /* HB_REPACKER_HH */ | 
|---|
| 411 |  | 
|---|