From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from lists.gentoo.org (pigeon.gentoo.org [208.92.234.80]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits)) (No client certificate requested) by finch.gentoo.org (Postfix) with ESMTPS id 5E0EF15800C for ; Fri, 30 Dec 2022 20:22:57 +0000 (UTC) Received: from pigeon.gentoo.org (localhost [127.0.0.1]) by pigeon.gentoo.org (Postfix) with SMTP id 7D55DE07A9; Fri, 30 Dec 2022 20:22:56 +0000 (UTC) Received: from smtp.gentoo.org (smtp.gentoo.org [IPv6:2001:470:ea4a:1:5054:ff:fec7:86e4]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (4096 bits) server-digest SHA256) (No client certificate requested) by pigeon.gentoo.org (Postfix) with ESMTPS id 547A4E079E for ; Fri, 30 Dec 2022 20:22:56 +0000 (UTC) Received: from oystercatcher.gentoo.org (oystercatcher.gentoo.org [148.251.78.52]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (4096 bits) server-digest SHA256) (No client certificate requested) by smtp.gentoo.org (Postfix) with ESMTPS id 30EDB340ADB for ; Fri, 30 Dec 2022 20:22:55 +0000 (UTC) Received: from localhost.localdomain (localhost [IPv6:::1]) by oystercatcher.gentoo.org (Postfix) with ESMTP id 87E9D7E7 for ; Fri, 30 Dec 2022 20:22:53 +0000 (UTC) From: "Sam James" To: gentoo-commits@lists.gentoo.org Content-Transfer-Encoding: 8bit Content-type: text/plain; charset=UTF-8 Reply-To: gentoo-dev@lists.gentoo.org, "Sam James" Message-ID: <1672431672.ab2870232ca91c7537058c64caf2012d0a2945d8.sam@gentoo> Subject: [gentoo-commits] repo/gentoo:master commit in: dev-db/mongodb/files/, dev-db/mongodb/ X-VCS-Repository: repo/gentoo X-VCS-Files: dev-db/mongodb/files/mongodb-4.4.10-boost-1.81.patch dev-db/mongodb/mongodb-4.4.10-r1.ebuild dev-db/mongodb/mongodb-5.0.5-r2.ebuild X-VCS-Directories: dev-db/mongodb/ dev-db/mongodb/files/ X-VCS-Committer: sam X-VCS-Committer-Name: Sam James X-VCS-Revision: ab2870232ca91c7537058c64caf2012d0a2945d8 X-VCS-Branch: master Date: Fri, 30 Dec 2022 20:22:53 +0000 (UTC) Precedence: bulk List-Post: List-Help: List-Unsubscribe: List-Subscribe: List-Id: Gentoo Linux mail X-BeenThere: gentoo-commits@lists.gentoo.org X-Auto-Response-Suppress: DR, RN, NRN, OOF, AutoReply X-Archives-Salt: df0c566b-650a-47b6-8806-6f2dcc369375 X-Archives-Hash: 2fd11a399b9261afc379979796191bfb commit: ab2870232ca91c7537058c64caf2012d0a2945d8 Author: Sam James gentoo org> AuthorDate: Fri Dec 30 20:19:15 2022 +0000 Commit: Sam James gentoo org> CommitDate: Fri Dec 30 20:21:12 2022 +0000 URL: https://gitweb.gentoo.org/repo/gentoo.git/commit/?id=ab287023 dev-db/mongodb: fix build w/ boost 1.81 Closes: https://bugs.gentoo.org/887037 Signed-off-by: Sam James gentoo.org> .../mongodb/files/mongodb-4.4.10-boost-1.81.patch | 317 +++++++++++++++++++++ dev-db/mongodb/mongodb-4.4.10-r1.ebuild | 1 + dev-db/mongodb/mongodb-5.0.5-r2.ebuild | 1 + 3 files changed, 319 insertions(+) diff --git a/dev-db/mongodb/files/mongodb-4.4.10-boost-1.81.patch b/dev-db/mongodb/files/mongodb-4.4.10-boost-1.81.patch new file mode 100644 index 000000000000..331f0c5b922a --- /dev/null +++ b/dev-db/mongodb/files/mongodb-4.4.10-boost-1.81.patch @@ -0,0 +1,317 @@ +https://bugs.gentoo.org/887037 + +Workaround https://github.com/boostorg/container/commit/99091420ae553b27345e04279fd19fe24fb684c1 +in Boost 1.81. + +Upstream s2 (as in real upstream, not MongoDB) has deviated substantially +from the version vendored. +--- a/src/third_party/s2/base/stl_decl_msvc.h ++++ b/src/third_party/s2/base/stl_decl_msvc.h +@@ -118,8 +118,8 @@ namespace msvchash { + class hash_multimap; + } // end namespace msvchash + +-using msvchash::hash_set; +-using msvchash::hash_map; ++using msvchash::hash_set = my_hash_set; ++using msvchash::hash_map = my_hash_map; + using msvchash::hash; + using msvchash::hash_multimap; + using msvchash::hash_multiset; +--- a/src/third_party/s2/base/stl_decl_osx.h ++++ b/src/third_party/s2/base/stl_decl_osx.h +@@ -68,8 +68,8 @@ using std::string; + + using namespace std; + using __gnu_cxx::hash; +-using __gnu_cxx::hash_set; +-using __gnu_cxx::hash_map; ++using __gnu_cxx::hash_set = my_hash_set; ++using __gnu_cxx::hash_map = my_hash_map; + using __gnu_cxx::select1st; + + /* On Linux (and gdrive on OSX), this comes from places like +--- a/src/third_party/s2/hash.h ++++ b/src/third_party/s2/hash.h +@@ -2,10 +2,10 @@ + #define THIRD_PARTY_S2_HASH_H_ + + #include +-#define hash_map std::unordered_map ++#define my_hash_map std::unordered_map + + #include +-#define hash_set std::unordered_set ++#define my_hash_set std::unordered_set + + #define HASH_NAMESPACE_START namespace std { + #define HASH_NAMESPACE_END } +--- a/src/third_party/s2/s2_test.cc ++++ b/src/third_party/s2/s2_test.cc +@@ -10,7 +10,7 @@ using std::reverse; + + #include + #include +-using __gnu_cxx::hash_set; ++using __gnu_cxx::hash_set = my_hash_map; + + #include "s2.h" + #include "base/logging.h" +@@ -709,8 +709,8 @@ TEST(S2, Frames) { + #if 0 + TEST(S2, S2PointHashSpreads) { + int kTestPoints = 1 << 16; +- hash_set set; +- hash_set points; ++ my_hash_set set; ++ my_hash_set points; + hash hasher; + S2Point base = S2Point(1, 1, 1); + for (int i = 0; i < kTestPoints; ++i) { +@@ -733,7 +733,7 @@ TEST(S2, S2PointHashCollapsesZero) { + double minus_zero = -zero; + EXPECT_NE(*reinterpret_cast(&zero), + *reinterpret_cast(&minus_zero)); +- hash_map map; ++ my_hash_map map; + S2Point zero_pt(zero, zero, zero); + S2Point minus_zero_pt(minus_zero, minus_zero, minus_zero); + +--- a/src/third_party/s2/s2cellid_test.cc ++++ b/src/third_party/s2/s2cellid_test.cc +@@ -10,7 +10,7 @@ using std::reverse; + + #include + #include +-using __gnu_cxx::hash_map; ++using __gnu_cxx::hash_map = my_hash_map; + + #include + #include +@@ -170,7 +170,7 @@ TEST(S2CellId, Tokens) { + static const int kMaxExpandLevel = 3; + + static void ExpandCell(S2CellId const& parent, vector* cells, +- hash_map* parent_map) { ++ my_hash_map* parent_map) { + cells->push_back(parent); + if (parent.level() == kMaxExpandLevel) return; + int i, j, orientation; +@@ -194,7 +194,7 @@ static void ExpandCell(S2CellId const& parent, vector* cells, + + TEST(S2CellId, Containment) { + // Test contains() and intersects(). +- hash_map parent_map; ++ my_hash_map parent_map; + vector cells; + for (int face = 0; face < 6; ++face) { + ExpandCell(S2CellId::FromFacePosLevel(face, 0, 0), &cells, &parent_map); +--- a/src/third_party/s2/s2loop.cc ++++ b/src/third_party/s2/s2loop.cc +@@ -120,7 +120,7 @@ bool S2Loop::IsValid(string* err) const { + } + } + // Loops are not allowed to have any duplicate vertices. +- hash_map vmap; ++ my_hash_map vmap; + for (int i = 0; i < num_vertices(); ++i) { + if (!vmap.insert(make_pair(vertex(i), i)).second) { + VLOG(2) << "Duplicate vertices: " << vmap[vertex(i)] << " and " << i; +--- a/src/third_party/s2/s2polygon.cc ++++ b/src/third_party/s2/s2polygon.cc +@@ -117,7 +117,7 @@ HASH_NAMESPACE_END + bool S2Polygon::IsValid(const vector& loops, string* err) { + // If a loop contains an edge AB, then no other loop may contain AB or BA. + if (loops.size() > 1) { +- hash_map > edges; ++ my_hash_map > edges; + for (size_t i = 0; i < loops.size(); ++i) { + S2Loop* lp = loops[i]; + for (int j = 0; j < lp->num_vertices(); ++j) { +--- a/src/third_party/s2/s2polygonbuilder.cc ++++ b/src/third_party/s2/s2polygonbuilder.cc +@@ -175,7 +175,7 @@ S2Loop* S2PolygonBuilder::AssembleLoop(S2Point const& v0, S2Point const& v1, + // This ensures that only CCW loops are constructed when possible. + + vector path; // The path so far. +- hash_map index; // Maps a vertex to its index in "path". ++ my_hash_map index; // Maps a vertex to its index in "path". + path.push_back(v0); + path.push_back(v1); + index[v1] = 1; +@@ -361,7 +361,7 @@ void S2PolygonBuilder::BuildMergeMap(PointIndex* index, MergeMap* merge_map) { + + // First, we build the set of all the distinct vertices in the input. + // We need to include the source and destination of every edge. +- hash_set vertices; ++ my_hash_set vertices; + for (EdgeSet::const_iterator i = edges_->begin(); i != edges_->end(); ++i) { + vertices.insert(i->first); + VertexSet const& vset = i->second; +@@ -370,7 +370,7 @@ void S2PolygonBuilder::BuildMergeMap(PointIndex* index, MergeMap* merge_map) { + } + + // Build a spatial index containing all the distinct vertices. +- for (hash_set::const_iterator i = vertices.begin(); ++ for (my_hash_set::const_iterator i = vertices.begin(); + i != vertices.end(); ++i) { + index->Insert(*i); + } +@@ -378,7 +378,7 @@ void S2PolygonBuilder::BuildMergeMap(PointIndex* index, MergeMap* merge_map) { + // Next, we loop through all the vertices and attempt to grow a maximial + // mergeable group starting from each vertex. + vector frontier, mergeable; +- for (hash_set::const_iterator vstart = vertices.begin(); ++ for (my_hash_set::const_iterator vstart = vertices.begin(); + vstart != vertices.end(); ++vstart) { + // Skip any vertices that have already been merged with another vertex. + if (merge_map->find(*vstart) != merge_map->end()) continue; +--- a/src/third_party/s2/s2polygonbuilder.h ++++ b/src/third_party/s2/s2polygonbuilder.h +@@ -262,7 +262,7 @@ class S2PolygonBuilder { + // current position to a new position, and also returns a spatial index + // containing all of the vertices that do not need to be moved. + class PointIndex; +- typedef hash_map MergeMap; ++ typedef my_hash_map MergeMap; + void BuildMergeMap(PointIndex* index, MergeMap* merge_map); + + // Moves a set of vertices from old to new positions. +@@ -282,7 +282,7 @@ class S2PolygonBuilder { + // once. We could have also used a multiset >, + // but this representation is a bit more convenient. + typedef multiset VertexSet; +- typedef hash_map EdgeSet; ++ typedef my_hash_map EdgeSet; + scoped_ptr edges_; + + // Unique collection of the starting (first) vertex of all edges, +--- a/src/third_party/s2/s2regioncoverer.cc ++++ b/src/third_party/s2/s2regioncoverer.cc +@@ -321,7 +321,7 @@ void S2RegionCoverer::GetInteriorCellUnion(S2Region const& region, + + void S2RegionCoverer::FloodFill( + S2Region const& region, S2CellId const& start, vector* output) { +- hash_set all; ++ my_hash_set all; + vector frontier; + output->clear(); + all.insert(start); +--- a/src/third_party/s2/s2regioncoverer_test.cc ++++ b/src/third_party/s2/s2regioncoverer_test.cc +@@ -11,7 +11,7 @@ using std::swap; + using std::reverse; + + #include +-using __gnu_cxx::hash_map; ++using __gnu_cxx::hash_map = my_hash_map; + + #include + using std::priority_queue; +@@ -65,7 +65,7 @@ static void CheckCovering(S2RegionCoverer const& coverer, + vector const& covering, + bool interior) { + // Keep track of how many cells have the same coverer.min_level() ancestor. +- hash_map min_level_cells; ++ my_hash_map min_level_cells; + for (int i = 0; i < covering.size(); ++i) { + int level = covering[i].level(); + EXPECT_GE(level, coverer.min_level()); +@@ -76,7 +76,7 @@ static void CheckCovering(S2RegionCoverer const& coverer, + if (covering.size() > coverer.max_cells()) { + // If the covering has more than the requested number of cells, then check + // that the cell count cannot be reduced by using the parent of some cell. +- for (hash_map::const_iterator i = min_level_cells.begin(); ++ for (my_hash_map::const_iterator i = min_level_cells.begin(); + i != min_level_cells.end(); ++i) { + EXPECT_EQ(i->second, 1); + } +--- a/src/third_party/s2/strings/split.cc ++++ b/src/third_party/s2/strings/split.cc +@@ -156,7 +156,7 @@ struct simple_insert_iterator { + // SplitStringToIterator{Using|AllowEmpty}(). + template + struct simple_hash_map_iterator { +- typedef hash_map hashmap; ++ typedef my_hash_map hashmap; + hashmap* t; + bool even; + typename hashmap::iterator curr; +@@ -246,8 +246,8 @@ void SplitStringAllowEmpty(const string& full, const char* delim, + } + + void SplitStringToHashsetAllowEmpty(const string& full, const char* delim, +- hash_set* result) { +- simple_insert_iterator > it(result); ++ my_hash_set* result) { ++ simple_insert_iterator > it(result); + SplitStringToIteratorAllowEmpty(full, delim, 0, it); + } + +@@ -258,7 +258,7 @@ void SplitStringToSetAllowEmpty(const string& full, const char* delim, + } + + void SplitStringToHashmapAllowEmpty(const string& full, const char* delim, +- hash_map* result) { ++ my_hash_map* result) { + simple_hash_map_iterator it(result); + SplitStringToIteratorAllowEmpty(full, delim, 0, it); + } +@@ -352,8 +352,8 @@ void SplitStringUsing(const string& full, + } + + void SplitStringToHashsetUsing(const string& full, const char* delim, +- hash_set* result) { +- simple_insert_iterator > it(result); ++ my_hash_set* result) { ++ simple_insert_iterator > it(result); + SplitStringToIteratorUsing(full, delim, it); + } + +@@ -364,7 +364,7 @@ void SplitStringToSetUsing(const string& full, const char* delim, + } + + void SplitStringToHashmapUsing(const string& full, const char* delim, +- hash_map* result) { ++ my_hash_map* result) { + simple_hash_map_iterator it(result); + SplitStringToIteratorUsing(full, delim, it); + } +--- a/src/third_party/s2/strings/split.h ++++ b/src/third_party/s2/strings/split.h +@@ -41,7 +41,7 @@ using namespace std; + void SplitStringAllowEmpty(const string& full, const char* delim, + vector* res); + void SplitStringToHashsetAllowEmpty(const string& full, const char* delim, +- hash_set* res); ++ my_hash_set* res); + void SplitStringToSetAllowEmpty(const string& full, const char* delim, + set* res); + // The even-positioned (0-based) components become the keys for the +@@ -50,7 +50,7 @@ void SplitStringToSetAllowEmpty(const string& full, const char* delim, + // if the key was already present in the hash table, or will be the + // empty string if the key is a newly inserted key. + void SplitStringToHashmapAllowEmpty(const string& full, const char* delim, +- hash_map* result); ++ my_hash_map* result); + + // ---------------------------------------------------------------------- + // SplitStringUsing() +@@ -66,7 +66,7 @@ void SplitStringToHashmapAllowEmpty(const string& full, const char* delim, + void SplitStringUsing(const string& full, const char* delim, + vector* res); + void SplitStringToHashsetUsing(const string& full, const char* delim, +- hash_set* res); ++ my_hash_set* res); + void SplitStringToSetUsing(const string& full, const char* delim, + set* res); + // The even-positioned (0-based) components become the keys for the +@@ -75,7 +75,7 @@ void SplitStringToSetUsing(const string& full, const char* delim, + // if the key was already present in the hash table, or will be the + // empty string if the key is a newly inserted key. + void SplitStringToHashmapUsing(const string& full, const char* delim, +- hash_map* result); ++ my_hash_map* result); + + // ---------------------------------------------------------------------- + // SplitOneIntToken() diff --git a/dev-db/mongodb/mongodb-4.4.10-r1.ebuild b/dev-db/mongodb/mongodb-4.4.10-r1.ebuild index b7037132f4ff..1c0661aee4ae 100644 --- a/dev-db/mongodb/mongodb-4.4.10-r1.ebuild +++ b/dev-db/mongodb/mongodb-4.4.10-r1.ebuild @@ -62,6 +62,7 @@ PATCHES=( "${FILESDIR}/${PN}-5.0.2-glibc-2.34.patch" "${FILESDIR}/${PN}-4.4.10-boost-1.79.patch" "${FILESDIR}/${PN}-4.4.10-no-force-lld.patch" + "${FILESDIR}/${PN}-4.4.10-boost-1.81.patch" ) S="${WORKDIR}/${MY_P}" diff --git a/dev-db/mongodb/mongodb-5.0.5-r2.ebuild b/dev-db/mongodb/mongodb-5.0.5-r2.ebuild index d0b26e5aafbf..d49a2e02c20b 100644 --- a/dev-db/mongodb/mongodb-5.0.5-r2.ebuild +++ b/dev-db/mongodb/mongodb-5.0.5-r2.ebuild @@ -69,6 +69,7 @@ PATCHES=( "${FILESDIR}/${PN}-5.0.2-skip-reqs-check.patch" "${FILESDIR}/${PN}-5.0.2-boost-1.79.patch" "${FILESDIR}/${PN}-5.0.5-no-force-lld.patch" + "${FILESDIR}/${PN}-4.4.10-boost-1.81.patch" ) S="${WORKDIR}/${MY_P}"