Compare commits

..

32 Commits

Author SHA1 Message Date
neru 7427357bc5 fix: dangling else
Build / build (push) Successful in 3m12s
2026-04-09 06:45:58 -03:00
neru 6b63d54eb8 build: modify tags, keep commits of failed runs
Build / build (push) Failing after 2m45s
2026-04-09 06:41:13 -03:00
neru a3df782245 feat: add getExePath 2026-04-09 06:40:34 -03:00
neru d4650aac84 feat: completely refactor dumper again
Build / build (push) Has been cancelled
2026-04-09 06:38:08 -03:00
neru 73b1c95257 chore: update items to reflect dumper changes 2026-04-09 06:37:59 -03:00
neru 93fe999249 feat: change items dump format 2026-04-09 06:37:53 -03:00
neru 393a591485 style: misc changes 2026-04-09 03:18:16 -03:00
neru 2b8cbafa19 feat: skip string scans for non bhvrdbd urls 2026-04-08 04:10:53 -03:00
neru 623542046a feat: add fun fact
Build / build (push) Successful in 3m19s
2026-04-07 19:43:20 -03:00
neru 90a6ab1ba8 fix: size_t to int casting
Build / build (push) Successful in 3m13s
2026-04-07 02:47:24 -03:00
neru 2c795a335f fix: typos
Build / build (push) Failing after 2m39s
2026-04-07 01:39:29 -03:00
neru 015a13eb86 fix: double assignment 2026-04-07 01:39:08 -03:00
neru b0524c0589 style: run clang-format 2026-04-07 01:37:09 -03:00
neru abd3c18040 style: remove leftover stuff 2026-04-07 01:34:29 -03:00
neru afb4a9c157 feat: give random item per req 2026-04-07 01:33:02 -03:00
neru 9c83499f0f feat: spoof bloodweb items 2026-04-07 01:28:03 -03:00
neru cd6d8df243 style: formatting 2026-04-07 01:04:20 -03:00
neru 64ded73f89 fix: spoof level on modifyCharacter to handle all requests 2026-04-07 01:04:15 -03:00
neru 022c67cab9 fix: initialize set only when used 2026-04-07 01:04:01 -03:00
neru cc07cd139c fix: leftover debug string 2026-04-07 01:03:43 -03:00
neru 490feb1be7 fix: killer detection logic 2026-04-07 01:03:19 -03:00
neru b5a8477ba2 fix: nurse not detected 2026-04-07 00:43:49 -03:00
neru e933515e45 fix: typo
Build / build (push) Successful in 3m10s
2026-04-06 21:54:57 -03:00
neru ff7284dfbf fix: update dumper to use xbox game 2026-04-06 21:54:52 -03:00
neru 7a710f2e17 feat: update catalog 2026-04-06 21:54:11 -03:00
neru b36d47a8fb chore: update file names to reflect platform 2026-04-06 21:51:49 -03:00
neru 6e2c709426 chore: add mappings
Build / build (push) Has been cancelled
2026-04-06 21:50:08 -03:00
neru 8929c0e39b fix: spoof level for every char < 15 or p0
Build / build (push) Successful in 4m27s
2026-04-02 01:21:03 -03:00
neru ab2385fc2f fix: comment out unused param
Build / build (push) Successful in 3m10s
2026-03-22 07:57:36 -03:00
neru 510241e311 fix: update to reflect callback changes
Build / build (push) Failing after 2m45s
2026-03-22 07:52:15 -03:00
neru 5e701e1e1b fix: make api res logs dbg only 2026-03-22 07:51:49 -03:00
neru edc4a8b500 feat: allow url modification 2026-03-22 07:41:48 -03:00
14 changed files with 599 additions and 546 deletions
+22 -20
View File
@@ -22,36 +22,31 @@ jobs:
if: ${{ github.event_name == 'push' }} if: ${{ github.event_name == 'push' }}
shell: bash shell: bash
run: | run: |
git fetch --tags || true git fetch --tags --force
# Determine last tag and log range if LAST_TAG=$(git describe --tags --abbrev=0 2>/dev/null); then
if git describe --tags --abbrev=0 >/dev/null 2>&1; then
LAST_TAG=$(git describe --tags --abbrev=0)
RANGE="${LAST_TAG}..HEAD" RANGE="${LAST_TAG}..HEAD"
else else
LAST_TAG="v0.1.9" LAST_TAG="v0.1.9"
RANGE="HEAD" RANGE="HEAD"
fi fi
# Calculate next version
IFS='.' read -r major minor patch <<< "${LAST_TAG#v}" IFS='.' read -r major minor patch <<< "${LAST_TAG#v}"
NEW_VERSION="$major.$minor.$((patch + 1))" NEW_VERSION="$major.$minor.$((patch + 1))"
NEW_TAG="v$NEW_VERSION"
# Log commits
CHANGELOG=$(git log $RANGE --oneline | sed 's/^/* /') CHANGELOG=$(git log $RANGE --oneline | sed 's/^/* /')
[ -z "$CHANGELOG" ] && CHANGELOG="Maintenance build." if [ -z "$CHANGELOG" ]; then
CHANGELOG="Re-run of version $NEW_VERSION or maintenance build."
fi
# Tag and push back to Gitea
git tag $NEW_TAG
git push origin $NEW_TAG
# Set outputs for next steps
echo "version-string=$NEW_VERSION" >> $GITHUB_OUTPUT echo "version-string=$NEW_VERSION" >> $GITHUB_OUTPUT
echo "changelog<<EOF" >> $GITHUB_OUTPUT echo "new-tag=v$NEW_VERSION" >> $GITHUB_OUTPUT
echo "## What's Changed" >> $GITHUB_OUTPUT {
echo "$CHANGELOG" >> $GITHUB_OUTPUT echo "changelog<<EOF"
echo "EOF" >> $GITHUB_OUTPUT echo "## What's Changed"
echo "$CHANGELOG"
echo "EOF"
} >> $GITHUB_OUTPUT
- name: Setup Tools Cache - name: Setup Tools Cache
id: tools-cache id: tools-cache
@@ -124,11 +119,18 @@ jobs:
name: unlocker-build name: unlocker-build
path: unlocker.zip path: unlocker.zip
- name: Finalize Version and Push Tag
if: ${{ github.event_name == 'push' && success() }}
run: |
git tag ${{ steps.calculate-version.outputs.new-tag }}
git push origin ${{ steps.calculate-version.outputs.new-tag }}
- name: Create Gitea Release - name: Create Gitea Release
if: ${{ github.event_name == 'push' }} if: ${{ github.event_name == 'push' && success() }}
uses: akkuman/gitea-release-action@v1 uses: akkuman/gitea-release-action@v1
with: with:
files: unlocker.zip files: unlocker.zip
tag_name: v${{ steps.calculate-version.outputs.version-string }} tag_name: ${{ steps.calculate-version.outputs.new-tag }}
name: Release v${{ steps.calculate-version.outputs.version-string }} name: Release ${{ steps.calculate-version.outputs.new-tag }}
body: ${{ steps.calculate-version.outputs.changelog }} body: ${{ steps.calculate-version.outputs.changelog }}
+1
View File
@@ -8,3 +8,4 @@ for /r "src" %%f in (*.cpp *.h *.cs) do (
echo done echo done
pause pause
+1 -1
View File
File diff suppressed because one or more lines are too long
+3 -7
View File
@@ -1,6 +1,5 @@
{ {
"Camper": { "Campers": [
"Items": [
"Father_Key_Card", "Father_Key_Card",
"FireflyLantern", "FireflyLantern",
"Item_Camper_AlexsToolbox", "Item_Camper_AlexsToolbox",
@@ -59,10 +58,8 @@
"Item_Survivor_MakeshiftFogVial", "Item_Survivor_MakeshiftFogVial",
"Item_Survivor_PrototypeFogVial", "Item_Survivor_PrototypeFogVial",
"Item_Survivor_VigosFogVial" "Item_Survivor_VigosFogVial"
] ],
}, "Slashers": [
"Slasher": {
"Powers": [
"Item_Blighted_Serum", "Item_Blighted_Serum",
"Item_Slasher_Beartrap", "Item_Slasher_Beartrap",
"Item_Slasher_Blinker", "Item_Slasher_Blinker",
@@ -108,4 +105,3 @@
"Item_Slasher_TormentMode" "Item_Slasher_TormentMode"
] ]
} }
}
Binary file not shown.
Binary file not shown.
+8 -8
View File
@@ -13,9 +13,9 @@ using Newtonsoft.Json;
class DumpByDaylight class DumpByDaylight
{ {
private const string _pakDir = "E:\\Program Files (x86)\\Steam\\steamapps\\common\\Dead by Daylight\\DeadByDaylight\\Content\\Paks"; private const string _pakDir = "D:\\XboxGames\\Dead By Daylight\\Content\\DeadByDaylight\\Content\\Paks";
private const string _aesKey = "0x22B1639B548124925CF7B9CBAA09F9AC295FCF0324586D6B37EE1D42670B39B3"; private const string _aesKey = "0x22B1639B548124925CF7B9CBAA09F9AC295FCF0324586D6B37EE1D42670B39B3";
private const string _mappingURL = "https://github.com/Masusder/FModel-DbdMappings/raw/refs/heads/main/Mappings/9.5.0/5.4.4-3172922+++DeadByDaylight+Quiche_REL-DeadByDaylight.usmap"; private const string _mappingURL = "https://git.neru.rip/neru/UnlockedByDaylight/raw/branch/main/res/mappings/latest-xbox.usmap";
public static async Task<string?> DownloadMappingFileAsync(string url, string savePath) public static async Task<string?> DownloadMappingFileAsync(string url, string savePath)
{ {
@@ -67,7 +67,7 @@ class DumpByDaylight
Console.WriteLine("\nProvider Initialized. Extracting Databases..."); Console.WriteLine("\nProvider Initialized. Extracting Databases...");
var dataPak = provider.GetArchive("pakchunk4-Windows.utoc"); var dataPak = provider.GetArchive("pakchunk4-WinGDK.utoc");
/* /*
* itemdb dump * itemdb dump
@@ -104,8 +104,8 @@ class DumpByDaylight
var itemsSerialized = new var itemsSerialized = new
{ {
Camper = new { Items = camperItems.OrderBy(x => x).ToList() }, Campers = camperItems.OrderBy(x => x).ToList(),
Slasher = new { Powers = slasherPowers.OrderBy(x => x).ToList() } Slashers = slasherPowers.OrderBy(x => x).ToList()
}; };
File.WriteAllText("items.json", JsonConvert.SerializeObject(itemsSerialized, Formatting.Indented)); File.WriteAllText("items.json", JsonConvert.SerializeObject(itemsSerialized, Formatting.Indented));
@@ -114,7 +114,7 @@ class DumpByDaylight
*/ */
searchPaths = dataPak.Files.Keys.Where(x => x.Contains($"/ItemAddonDB.uasset", StringComparison.OrdinalIgnoreCase)).ToList(); searchPaths = dataPak.Files.Keys.Where(x => x.Contains($"/ItemAddonDB.uasset", StringComparison.OrdinalIgnoreCase)).ToList();
var camperAddons = new List<string>(); var camperAddons = new List<string>();
var slaherAddons = new List<string>(); var slasherAddons = new List<string>();
foreach (var path in searchPaths) foreach (var path in searchPaths)
{ {
@@ -135,7 +135,7 @@ class DumpByDaylight
} }
if (isSlasherAddon) if (isSlasherAddon)
slaherAddons.Add(row.Key.Text); slasherAddons.Add(row.Key.Text);
else else
camperAddons.Add(row.Key.Text); camperAddons.Add(row.Key.Text);
} }
@@ -144,7 +144,7 @@ class DumpByDaylight
var addonsSerialized = new var addonsSerialized = new
{ {
Slashers = slaherAddons.OrderBy(x => x).ToList(), Slashers = slasherAddons.OrderBy(x => x).ToList(),
Campers = camperAddons.OrderBy(x => x).ToList() Campers = camperAddons.OrderBy(x => x).ToList()
}; };
File.WriteAllText("addons.json", JsonConvert.SerializeObject(addonsSerialized, Formatting.Indented)); File.WriteAllText("addons.json", JsonConvert.SerializeObject(addonsSerialized, Formatting.Indented));
-13
View File
@@ -5,18 +5,6 @@
#include <windows.h> #include <windows.h>
#include <wininet.h> #include <wininet.h>
//#include <fstream>
//#include <iostream>
//#include <string>
//#include <string_view>
//#include <vector>
//#include <format>
//#include <mutex>
//#include <ctime>
//#include <regex>
//#include <unordered_set>
//#include <simdjson.h>
bool setProxy(bool enable, const std::string& proxyAddr) bool setProxy(bool enable, const std::string& proxyAddr)
{ {
@@ -83,7 +71,6 @@ BOOL WINAPI consoleHandler(DWORD dwType)
running = false; running = false;
cleanup(); cleanup();
exit(0); exit(0);
//return TRUE;
} }
return FALSE; return FALSE;
} }
+5
View File
@@ -335,6 +335,9 @@ void Proxy::handleClient(SOCKET hClientSocket)
{ {
std::string emptyBody = ""; std::string emptyBody = "";
OnClientRequest.run(url, emptyBody, headers); OnClientRequest.run(url, emptyBody, headers);
if (!pendingUrls.empty()) pendingUrls.back() = url;
SSL_write(remoteSSL, headers.data(), (int)headers.size()); SSL_write(remoteSSL, headers.data(), (int)headers.size());
clientStream.buffer.erase(0, clientStream.headersEnd + 4); clientStream.buffer.erase(0, clientStream.headersEnd + 4);
clientStream.reset(); clientStream.reset();
@@ -383,6 +386,8 @@ void Proxy::handleClient(SOCKET hClientSocket)
if (complete) if (complete)
{ {
OnClientRequest.run(url, body, headers); OnClientRequest.run(url, body, headers);
if (!pendingUrls.empty() && pendingUrls.back() != url) pendingUrls.back() = url;
SSL_write(remoteSSL, headers.data(), (int)headers.size()); SSL_write(remoteSSL, headers.data(), (int)headers.size());
SSL_write(remoteSSL, clientStream.buffer.data() + bodyStart, SSL_write(remoteSSL, clientStream.buffer.data() + bodyStart,
(int)(clientStream.buffer.size() - bodyStart)); (int)(clientStream.buffer.size() - bodyStart));
+1 -1
View File
@@ -25,7 +25,7 @@ class Proxy
bool Init(); bool Init();
void Shutdown(); void Shutdown();
CallbackEvent<const std::string&, const std::string&, std::string&> OnClientRequest; CallbackEvent<std::string&, const std::string&, std::string&> OnClientRequest;
CallbackEvent<const std::string&, std::string&, std::string&> OnServerResponse; CallbackEvent<const std::string&, std::string&, std::string&> OnServerResponse;
private: private:
+365 -339
View File
@@ -1,10 +1,13 @@
#include "spoofing.h" #include "spoofing.h"
#include "utils.h"
#include <string>
#include <fstream> #include <fstream>
#include <algorithm> #include <random>
#include <unordered_set> #include <vector>
#include <regex>
#include <minwindef.h> #include <time.h>
#include <nerutils/log.h> #include <nerutils/log.h>
@@ -12,205 +15,295 @@
using json = nlohmann::json; using json = nlohmann::json;
std::string getExeDir() static std::random_device rd;
static std::mt19937 gen(rd());
/*
utils
*/
std::unordered_set<std::string> slasherNames = {
"Chuckles", "Bob", "HillBilly", "Nurse", "Shape", "Witch", "Killer07", "Cannibal", "Bear",
"Nightmare", "Pig", "Clown", "Spirit", "Plague", "Ghostface", "Demogorgon", "Oni", "Gunslinger"};
bool isSlasher(std::string name)
{ {
char buffer[MAX_PATH]; if (slasherNames.contains(name) || (name.length() == 3 && name[0] == 'K')) return true;
GetModuleFileNameA(NULL, buffer, MAX_PATH); return false;
std::string path(buffer);
size_t pos = path.find_last_of("\\/");
if (pos != std::string::npos) return path.substr(0, pos + 1);
return "";
} }
/*
setup / init
*/
void Spoofer::init(Proxy* proxy) void Spoofer::init(Proxy* proxy)
{ {
loadData();
registerListeners(proxy); registerListeners(proxy);
loadData();
} }
void Spoofer::registerListeners(Proxy* proxy) void Spoofer::registerListeners(Proxy* proxy)
{ {
Log::info("Registering Proxy event listeners for spoofer");
proxy->OnServerResponse.addListener([this](const std::string& url, std::string& body, std::string& respHeaders) { proxy->OnServerResponse.addListener([this](const std::string& url, std::string& body, std::string& respHeaders) {
this->serverResponseHandler(url, body, respHeaders); this->serverResponseHandler(url, body, respHeaders);
}); });
proxy->OnClientRequest.addListener( proxy->OnClientRequest.addListener([this](std::string& url, const std::string& body, std::string& reqHeaders) {
[this](const std::string& url, const std::string& body, std::string& reqHeaders) {
this->clientRequestHandler(url, body, reqHeaders); this->clientRequestHandler(url, body, reqHeaders);
}); });
} }
#define LOADDATA(fileName, type, camperSet, slasherSet, errorMsg) \
std::string path##type = utils::getExePath() + fileName; \
std::ifstream file##type(path##type); \
if (file##type.is_open()) \
{ \
std::stringstream buff; \
buff << file##type.rdbuf(); \
if (!parseStackable(buff.str(), camperSet, slasherSet)) \
Log::error("Failed to parse {} - {}", fileName, errorMsg); \
else \
Log::verbose("Loaded {} camper ids and {} slasher ids from {}", camperSet.size(), slasherSet.size(), \
fileName); \
} \
else \
Log::warning("Missing {} - {}", fileName, errorMsg);
void Spoofer::loadData() void Spoofer::loadData()
{ {
Log::info("Loading dump data"); Log::info("Loading data");
std::string catalogDumpPath = getExeDir() + "catalog.json"; std::string catalogPath = utils::getExePath() + "catalog.json";
std::ifstream catalogFile(catalogDumpPath); std::ifstream catalogFile(catalogPath);
if (catalogFile.is_open()) if (catalogFile.is_open())
{ {
std::stringstream buffer; std::stringstream buff;
buffer << catalogFile.rdbuf(); buff << catalogFile.rdbuf();
parseCatalog(buffer.str()); if (!parseCatalog(buff.str())) Log::warning("Failed to parse catalog.json - Customizations won't be unlocked");
} }
else else
Log::warning("Missing catalog.json"); Log::warning("Missing catalog.json - Customizations won't be unlocked");
std::string itemDumpPath = getExeDir() + "items.json"; LOADDATA("items.json", Items, _camperItemIds, _slasherPowerIds, "Items won't be added");
std::ifstream itemFile(itemDumpPath); LOADDATA("offerings.json", Offerings, _camperOfferingIds, _slasherOfferingIds, "Offerings won't be added");
if (itemFile.is_open()) LOADDATA("addons.json", Addons, _camperAddonIds, _slasherAddonIds, "Addons won't be added");
{ LOADDATA("perks.json", Perks, _camperPerkIds, _slasherPerkIds, "Perks won't be added");
try }
{
std::string content((std::istreambuf_iterator<char>(itemFile)), std::istreambuf_iterator<char>());
json doc = json::parse(content);
if (doc.contains("Camper") && doc["Camper"].is_object()) /*
if (doc["Camper"].contains("Items") && doc["Camper"]["Items"].is_array()) data parsing
for (const auto& item : doc["Camper"]["Items"]) */
if (item.is_string()) _camperItemIds.insert(item.get<std::string>()); bool Spoofer::parseCatalog(std::string data)
if (doc.contains("Slasher") && doc["Slasher"].is_object())
if (doc["Slasher"].contains("Powers") && doc["Slasher"]["Powers"].is_array())
for (const auto& item : doc["Slasher"]["Powers"])
if (item.is_string()) _slasherPowerIds.insert(item.get<std::string>());
}
catch (const json::parse_error& e)
{ {
Log::error("JSON parse error in {}: {}", "items.json", e.what()); json doc = json::parse(data, nullptr, false);
}
}
else
Log::warning("Missing items.json");
std::string offeringDumpPath = getExeDir() + "offerings.json"; if (doc.is_discarded())
std::ifstream offeringsFile(offeringDumpPath);
if (offeringsFile.is_open())
{ {
try Log::error("Failed to parse catalog");
{ return false;
std::string content((std::istreambuf_iterator<char>(offeringsFile)), std::istreambuf_iterator<char>());
json doc = json::parse(content);
if (doc.contains("Slashers") && doc["Slashers"].is_array())
for (const auto& offering : doc["Slashers"])
if (offering.is_string()) _slasherOfferingIds.insert(offering.get<std::string>());
if (doc.contains("Campers") && doc["Campers"].is_array())
for (const auto& offering : doc["Campers"])
if (offering.is_string()) _camperOfferingIds.insert(offering.get<std::string>());
} }
catch (const json::parse_error& e)
{
Log::error("JSON parse error in {}: {}", "offerings.json", e.what());
}
}
else
Log::warning("Missing offerings.json");
std::string addonDumpPath = getExeDir() + "addons.json";
std::ifstream addonFile(addonDumpPath);
if (addonFile.is_open())
{
try
{
std::string content((std::istreambuf_iterator<char>(addonFile)), std::istreambuf_iterator<char>());
json doc = json::parse(content);
if (doc.contains("Slashers") && doc["Slashers"].is_array())
for (const auto& item : doc["Slashers"])
if (item.is_string()) _slasherAddonIds.insert(item.get<std::string>());
if (doc.contains("Campers") && doc["Campers"].is_array())
for (const auto& power : doc["Campers"])
if (power.is_string()) _camperAddonIds.insert(power.get<std::string>());
}
catch (const json::parse_error& e)
{
Log::error("JSON parse error in {}: {}", "addons.json", e.what());
}
}
else
Log::warning("Missing addons.json");
try try
{ {
std::string perkDumpPaths = getExeDir() + "perks.json"; const auto& catalogData = doc.at("data");
std::ifstream perkFile(perkDumpPaths);
if (perkFile.is_open()) auto extractIds = [&](const std::string& key, std::unordered_set<std::string>& targetSet) {
if (catalogData.contains(key) && catalogData[key].contains("items"))
for (const auto& id : catalogData[key]["items"])
{ {
std::string content((std::istreambuf_iterator<char>(perkFile)), std::istreambuf_iterator<char>()); if (id.is_string()) targetSet.insert(id.get<std::string>());
json doc = json::parse(content);
if (doc.contains("Slashers") && doc["Slashers"].is_array())
for (const auto& item : doc["Slashers"])
if (item.is_string()) _slasherPerkIds.insert(item.get<std::string>());
if (doc.contains("Campers") && doc["Campers"].is_array())
for (const auto& power : doc["Campers"])
if (power.is_string()) _camperPerkIds.insert(power.get<std::string>());
} }
else else
Log::warning("Missing perks.json"); Log::warning("Catalog missing or invalid category: {}", key);
};
extractIds("item", _catalogItemIds);
extractIds("outfit", _catalogOutfitIds);
} }
catch (const json::parse_error& e) catch (const json::exception& e)
{ {
Log::error("JSON parse error in {}: {}", "perks.json", e.what()); Log::error("Invalid catalog format: {}", e.what());
return false;
}
Log::info("Parsed {} items and {} outfits from catalog", _catalogItemIds.size(), _catalogOutfitIds.size());
return true;
}
bool Spoofer::parseStackable(std::string data, std::unordered_set<std::string>& camperSet,
std::unordered_set<std::string>& slasherSet)
{
json doc = json::parse(data, nullptr, false);
if (doc.is_discarded())
{
Log::error("JSON parse error @ parseStackable");
return false;
}
auto populate = [&](const std::string& key, std::unordered_set<std::string>& targetSet) {
if (doc.contains(key) && doc[key].is_array())
for (const auto& item : doc[key])
{
if (item.is_string()) targetSet.insert(item.get<std::string>());
}
else
Log::warning("Missing stackables array ({})", key);
};
populate("Slashers", slasherSet);
populate("Campers", camperSet);
return true;
}
/*
misc functions
*/
std::string Spoofer::getRandomItem()
{
std::vector<const std::unordered_set<std::string>*> allSets = {
&_camperItemIds, &_slasherPowerIds, &_camperOfferingIds, &_slasherOfferingIds,
&_camperAddonIds, &_slasherAddonIds, &_camperPerkIds, &_slasherPerkIds};
std::vector<const std::unordered_set<std::string>*> validSets;
for (auto* s : allSets)
if (!s->empty()) validSets.push_back(s);
if (validSets.empty()) return "Spring2024Offering";
std::uniform_int_distribution<> setDist(0, static_cast<int>(validSets.size()) - 1);
const auto& selectedSet = *validSets[setDist(gen)];
std::uniform_int_distribution<> itemDist(0, static_cast<int>(selectedSet.size()) - 1);
auto it = selectedSet.begin();
std::advance(it, itemDist(gen));
return *it;
}
int Spoofer::getRandomQuantity()
{
std::uniform_int_distribution<> dist(64, 97);
return dist(gen);
}
void Spoofer::generateBloodweb(nlohmann::json& js)
{
if (!js.is_object()) js = json::object();
std::vector<std::string> paths;
json ringDataArray = json::array();
ringDataArray.push_back({{"nodeData", {{{"nodeId", "0"}, {"state", "Collected"}}}}});
int nodesPerRing[] = {6, 12, 12};
std::vector<std::string> prevRingNodes = {"0"};
for (int ring = 1; ring <= 3; ++ring)
{
json nodeDataArray = json::array();
std::vector<std::string> currentRingNodes;
int numNodes = nodesPerRing[ring - 1];
for (int i = 1; i <= numNodes; ++i)
{
std::string childId = std::to_string((ring * 100) + i);
currentRingNodes.push_back(childId);
int parentIndex = (i - 1) / (numNodes / static_cast<int>(prevRingNodes.size()));
std::string parentId = prevRingNodes[(std::min)(parentIndex, (int)prevRingNodes.size() - 1)];
paths.push_back(parentId + "_" + childId);
std::string item = getRandomItem();
if (item.empty()) item = "Spring2024Offering";
nodeDataArray.push_back({{"nodeId", childId}, {"state", "Collected"}, {"contentId", item}});
}
ringDataArray.push_back({{"nodeData", nodeDataArray}});
prevRingNodes = std::move(currentRingNodes);
}
js["paths"] = paths;
js["ringData"] = ringDataArray;
}
/*
data modification
*/
void Spoofer::modifyCharacterData(json& js)
{
if (!js.contains("characterName") || !js["characterName"].is_string())
{
Log::verbose("attempted to modify invalid char");
return; return;
} }
Log::info("Loaded: Camper items={}, Slasher powers={}, " std::string name = js["characterName"];
"Slasher addons={}, Camper addons={}, Camper perks={}, Slasher perks={}" bool slasher = isSlasher(js["characterName"]);
"Slasher offerings={}, Camper offerings={}",
_camperItemIds.size(), _slasherPowerIds.size(), _slasherAddonIds.size(), _camperAddonIds.size(),
_camperPerkIds.size(), _slasherPerkIds.size(), _slasherOfferingIds.size(), _camperOfferingIds.size());
}
void Spoofer::parseCatalog(std::string data) bool needsSpoofing = false;
if (js.value("isEntitled", true) == false)
{ {
try _unownedCharacters.insert(name);
js["isEntitled"] = true;
js["purchaseInfo"] = {{"quantity", 1},
{"origin", "PlayerInventory"},
{"reason", "Item(s) added via Purchase"},
{"lastUpdateAt", std::time(nullptr)},
{"objectId", name}};
needsSpoofing = true;
}
else if (_unownedCharacters.contains(name))
needsSpoofing = true;
/*
modifications for unowned characters (spoof level and fake bloodweb)
*/
if (needsSpoofing)
{ {
json doc = json::parse(data); if (js.contains("bloodWebLevel") && js["bloodWebLevel"].is_number() && js["bloodWebLevel"] <= 15)
if (!js.contains("prestigeLevel") || (js["prestigeLevel"].is_number() && js["prestigeLevel"] <= 0))
js["bloodWebLevel"] = 16;
if (doc.contains("data")) doc = doc["data"]; if (js.contains("bloodWebData")) generateBloodweb(js["bloodWebData"]);
if (doc.contains("item") && doc["item"].contains("items") && doc["item"]["items"].is_array())
for (const auto& item : doc["item"]["items"])
if (item.is_string()) _catalogItemIds.insert(item.get<std::string>());
if (doc.contains("outfit") && doc["outfit"].contains("items") && doc["outfit"]["items"].is_array())
for (const auto& item : doc["outfit"]["items"])
if (item.is_string()) _catalogOutfitIds.insert(item.get<std::string>());
Log::info("Parsed {} items and {} outfits from catalog", _catalogItemIds.size(), _catalogOutfitIds.size());
} }
catch (const json::parse_error& e) else
{ {
Log::error("JSON parse error in {}: {}", "parseCatalog", e.what()); /*
} ghost node hotfix (untested)
} */
/* if (js.contains("bloodWebData") && js["bloodWebData"].contains("ringData"))
void Spoofer::parseAndDumpCatalog(std::string& data)
{ {
std::string path = getExeDir() + "catalog.json"; auto& ringData = js["bloodWebData"]["ringData"];
std::ofstream file(path);
file << data;
file.close();
Log::info("Raw catalog saved to {}", path);
parseCatalog(data); for (auto& ring : ringData)
{
if (ring.contains("nodeData") && ring["nodeData"].is_array())
{
for (auto& node : ring["nodeData"])
{
if (node.contains("nodeId") && node["nodeId"] != "0")
{
if (!node.contains("contentId") || node["contentId"].get<std::string>().empty())
{
node["contentId"] = "Spring2024Offering";
if (!node.contains("state"))
node["state"] = "Available";
}
}
}
}
}
}*/
} }
void Spoofer::modifyCharacterData(json& js) /*
item spoofing
*/
if (js.contains("characterItems") && js["characterItems"].is_array())
{ {
std::unordered_set<std::string> existingItemIds; std::unordered_set<std::string> existingItemIds;
bool isSlasher = false;
if (js.contains("characterName") && js["characterName"].is_string())
{
std::string name = js["characterName"];
if (name == "Bear" || name == "Bob" || name == "Chuckles" || (name.length() >= 2 && name[0] == 'K'))
isSlasher = true;
}
std::unordered_set<std::string> stackableIds; std::unordered_set<std::string> stackableIds;
stackableIds.insert(_camperItemIds.begin(), _camperItemIds.end()); stackableIds.insert(_camperItemIds.begin(), _camperItemIds.end());
@@ -219,256 +312,189 @@ void Spoofer::modifyCharacterData(json& js)
stackableIds.insert(_slasherAddonIds.begin(), _slasherAddonIds.end()); stackableIds.insert(_slasherAddonIds.begin(), _slasherAddonIds.end());
stackableIds.insert(_slasherOfferingIds.begin(), _slasherOfferingIds.end()); stackableIds.insert(_slasherOfferingIds.begin(), _slasherOfferingIds.end());
if (js.contains("characterItems") && js["characterItems"].is_array())
{
for (auto& item : js["characterItems"]) for (auto& item : js["characterItems"])
{ {
/*
set existing items to rnd number
*/
if (item.contains("itemId") && item["itemId"].is_string()) if (item.contains("itemId") && item["itemId"].is_string())
{ {
std::string itemId = item["itemId"]; std::string itemId = item["itemId"];
existingItemIds.insert(itemId); existingItemIds.insert(itemId);
if (stackableIds.contains(itemId)) item["quantity"] = getRandomQuantity();
if (_slasherPowerIds.find(itemId) != _slasherPowerIds.end())
isSlasher = true;
else if (stackableIds.contains(itemId))
item["quantity"] = 100;
} }
} }
if (!isSlasher) auto appendItems = [&](const std::unordered_set<std::string>& idList, bool isPerk) {
for (const auto& itemId : idList)
if (existingItemIds.find(itemId) == existingItemIds.end())
js["characterItems"].push_back(
{{"itemId", itemId}, {"quantity", isPerk ? 3 : getRandomQuantity()}});
};
if (!slasher)
{ {
for (const std::string& itemId : _camperItemIds) appendItems(_camperItemIds, false);
if (existingItemIds.find(itemId) == existingItemIds.end()) appendItems(_camperAddonIds, false);
js["characterItems"].push_back({{"itemId", itemId}, {"quantity", 100}}); appendItems(_camperOfferingIds, false);
for (const std::string& itemId : _camperAddonIds) appendItems(_camperPerkIds, true);
if (existingItemIds.find(itemId) == existingItemIds.end())
js["characterItems"].push_back({{"itemId", itemId}, {"quantity", 100}});
for (const std::string& itemId : _camperOfferingIds)
if (existingItemIds.find(itemId) == existingItemIds.end())
js["characterItems"].push_back({{"itemId", itemId}, {"quantity", 100}});
for (const std::string& itemId : _camperPerkIds)
if (existingItemIds.find(itemId) == existingItemIds.end())
js["characterItems"].push_back({{"itemId", itemId}, {"quantity", 3}});
} }
else else
{ {
for (const std::string& itemId : _slasherAddonIds) appendItems(_slasherAddonIds, false);
if (existingItemIds.find(itemId) == existingItemIds.end()) appendItems(_slasherOfferingIds, false);
js["characterItems"].push_back({{"itemId", itemId}, {"quantity", 100}}); appendItems(_slasherPerkIds, true);
for (const std::string& itemId : _slasherOfferingIds)
if (existingItemIds.find(itemId) == existingItemIds.end())
js["characterItems"].push_back({{"itemId", itemId}, {"quantity", 100}});
for (const std::string& itemId : _slasherPerkIds)
if (existingItemIds.find(itemId) == existingItemIds.end())
js["characterItems"].push_back({{"itemId", itemId}, {"quantity", 3}});
}
} }
} }
void Spoofer::serverResponseHandler(const std::string& url, std::string& body, std::string& respHeaders) Log::verbose("Spoofed data for character {}", name);
{ }
if (url.find("bhvrdbd.com") != std::string::npos) Log::verbose("BHVR api res @ {}", url);
if (url.find("api/v1/extensions/store/getCatalogItems") != std::string::npos) return parseAndDumpCatalog(body); /*
endpoint handlers
*/
void Spoofer::onGetAll(std::string& body)
{
json doc = json::parse(body, nullptr, false);
if (doc.is_discarded()) return Log::error("JSON parse error for dbd-character-data/get-all");
if (!doc.contains("list") || !doc["list"].is_array())
return Log::error("Invalid json for dbd-character-data/get-all");
if (url.find("api/v1/dbd-inventories/all") != std::string::npos) for (auto& charData : doc["list"])
modifyCharacterData(charData);
body = doc.dump();
}
void Spoofer::onInventoryAll(std::string& body)
{ {
try json doc = json::parse(body, nullptr, false);
{ if (doc.is_discarded()) return Log::error("JSON parse error for dbd-inventories/all");
json doc = json::parse(body); if (!doc.contains("inventoryItems") || !doc["inventoryItems"].is_array())
return Log::error("Invalid json for JSON parse error for dbd-inventories/all");
auto& itemsArr = doc["inventoryItems"]; auto& itemsArr = doc["inventoryItems"];
std::unordered_set<std::string> existingIds;
int64_t now = std::time(nullptr);
std::unordered_set<std::string> foundObjects; /*
std::unordered_set<std::string> foundAddons; item updates
std::unordered_set<std::string> foundPerks; */
std::unordered_set<std::string> foundOfferings; struct Category
std::unordered_set<std::string> foundCatalogItems; {
const std::unordered_set<std::string>& source;
int quantity;
};
std::unordered_set<std::string> offeringIds; std::vector<Category> categories = {{_camperPerkIds, 3}, {_slasherPerkIds, 3}, {_camperOfferingIds, -1},
offeringIds.insert(_camperOfferingIds.begin(), _camperOfferingIds.end()); {_slasherOfferingIds, -1}, {_catalogOutfitIds, 1}, {_catalogItemIds, 1}};
offeringIds.insert(_slasherOfferingIds.begin(), _slasherOfferingIds.end());
std::unordered_set<std::string> catalogIds;
catalogIds.insert(_catalogOutfitIds.begin(), _catalogOutfitIds.end());
catalogIds.insert(_catalogItemIds.begin(), _catalogItemIds.end());
std::unordered_set<std::string> perkIds;
perkIds.insert(_slasherPerkIds.begin(), _slasherPerkIds.end());
perkIds.insert(_camperPerkIds.begin(), _camperPerkIds.end());
for (auto& item : itemsArr) for (auto& item : itemsArr)
{ {
std::string objectId = item["objectId"]; std::string id = item.value("objectId", "");
if (id.empty()) continue;
if (perkIds.find(objectId) != perkIds.end()) existingIds.insert(id);
for (const auto& cat : categories)
{ {
foundPerks.insert(objectId); if (cat.source.contains(id))
item["quantity"] = 3; {
continue; item["quantity"] = (cat.quantity == -1) ? getRandomQuantity() : cat.quantity;
break;
} }
if (offeringIds.find(objectId) != offeringIds.end())
{
foundOfferings.insert(objectId);
item["quantity"] = 100;
continue;
}
if (catalogIds.find(objectId) != catalogIds.end())
{
foundCatalogItems.insert(objectId);
continue;
} }
} }
for (const std::string& id : perkIds) /*
item inserts
*/
for (const auto& cat : categories)
{ {
if (foundPerks.find(id) == foundPerks.end()) for (const std::string& id : cat.source)
{ {
itemsArr.push_back({ if (!existingIds.contains(id))
{"objectId", id}, {
{"quantity", 3}, int qty = (cat.quantity == -1) ? getRandomQuantity() : cat.quantity;
{"lastUpdateAt", std::time(nullptr)}, itemsArr.push_back({{"objectId", id}, {"quantity", qty}, {"lastUpdateAt", now}});
}); existingIds.insert(id);
}
} }
} }
for (const std::string& id : offeringIds)
{
if (foundOfferings.find(id) == foundOfferings.end())
{
itemsArr.push_back({
{"objectId", id},
{"quantity", 100},
{"lastUpdateAt", std::time(nullptr)},
});
}
}
for (const std::string& id : catalogIds)
{
if (foundCatalogItems.find(id) == foundCatalogItems.end())
{
itemsArr.push_back({
{"objectId", id},
{"quantity", 1},
{"lastUpdateAt", std::time(nullptr)},
});
}
}
std::string updatedJson = doc.dump();
body = updatedJson;
Log::verbose("Inventory updated: Items={}, Addons={}, Perks={}, Offerings={}", foundObjects.size(),
foundAddons.size(), foundPerks.size(), foundOfferings.size());
return;
}
catch (const json::parse_error& e)
{
Log::error("JSON parse error in {}: {}", url, e.what());
return;
}
return;
}
if (url.find("api/v1/dbd-character-data/get-all") != std::string::npos)
{
try
{
json doc = json::parse(body);
auto& charList = doc["list"];
for (auto& charInfo : charList)
{
if (charInfo["isEntitled"] == false)
{
charInfo["isEntitled"] = true;
if (charInfo.contains("bloodWebLevel")) charInfo["bloodWebLevel"] = 15;
if (charInfo.contains("bloodWebData")) charInfo["bloodWebData"]["level"] = 1;
if (charInfo.contains("origin"))
{
charInfo["purchaseInfo"] = {{"quantity", 1},
{"origin", "PlayerInventory"},
{"reason", "Item(s) added via Purchase"},
{"lastUpdateAt", 1770702482},
{"objectId", charInfo["characterName"]}};
}
}
modifyCharacterData(charInfo);
}
body = doc.dump(); body = doc.dump();
} Log::verbose("Inventory spoofed");
catch (const json::parse_error& e)
{
Log::error("JSON parse error in {}: {}", url, e.what());
}
return;
} }
if (url.find("api/v1/dbd-character-data/") != std::string::npos) void Spoofer::onBloodweb(std::string& body, std::string& respHeaders)
{ {
try json doc = json::parse(body, nullptr, false);
{ if (doc.is_discarded()) return Log::error("JSON parse error for bloodweb response");
json doc = json::parse(body);
if (body.find("NotAllowedException") != std::string::npos && body.find("not owned") != std::string::npos) if (body.find("NotAllowedException") != std::string::npos && body.find("not owned") != std::string::npos)
{ {
Log::info("Spoofing bloodweb error error for unowned character"); Log::info("Spoofing bloodweb error for unowned character");
json mock; json mock;
mock["bloodWebLevelChanged"] = false; mock["bloodWebLevelChanged"] = false;
mock["updatedWallets"] = json::array(); mock["updatedWallets"] = json::array();
mock["bloodWebLevel"] = 15; mock["bloodWebLevel"] = 16;
mock["prestigeLevel"] = 0; mock["prestigeLevel"] = 0;
mock["bloodWebData"] = {{"ringData", json::array()}, {"paths", json::array()}}; mock["bloodWebData"] = json::object();
mock["characterItems"] = json::array(); mock["characterItems"] = json::array();
mock["characterName"] = this->_lastBloodWebChar; mock["characterName"] = this->_lastBloodWebChar;
mock["isEntitled"] = true;
mock["purchaseInfo"] = {{"quantity", 1},
{"origin", "PlayerInventory"},
{"reason", "Item(s) added via Purchase"},
{"lastUpdateAt", std::time(nullptr)},
{"objectId", this->_lastBloodWebChar}};
_unownedCharacters.insert(this->_lastBloodWebChar); // probably not needed but just in case
modifyCharacterData(mock); modifyCharacterData(mock);
size_t firstSpace = respHeaders.find(' '); std::regex statusRegex(R"(HTTP\/\d\.\d\s+403)");
if (firstSpace != std::string::npos) respHeaders = std::regex_replace(respHeaders, statusRegex, "HTTP/1.1 200");
{
respHeaders.replace(firstSpace + 1, 3, "200");
}
body = mock.dump(); body = mock.dump();
Log::verbose("Spoofed bloodweb request for unowned character.");
return; return;
} }
else
{
modifyCharacterData(doc);
body = doc.dump();
}
return;
}
catch (const json::parse_error& e)
{
Log::error("JSON parse error in {}: {}", url, e.what());
}
return;
}
}
void Spoofer::clientRequestHandler(const std::string& url, const std::string& body, std::string& /*reqHeaders*/) modifyCharacterData(doc);
body = doc.dump();
Log::verbose("Spoofed bloodweb items for owned character");
}
/*
event handlers
*/
void Spoofer::serverResponseHandler(const std::string& url, std::string& body, std::string& respHeaders)
{ {
if (url.find("bhvrdbd.com") == std::string::npos) return;
std::lock_guard<std::mutex> lock(_mtx);
#ifdef _DEBUG
Log::verbose("BHVR api res @ {}", url);
#endif
if (url.find("api/v1/dbd-character-data/get-all") != std::string::npos) return onGetAll(body);
if (url.find("api/v1/dbd-inventories/all") != std::string::npos) return onInventoryAll(body);
if (url.find("api/v1/dbd-character-data/bloodweb") != std::string::npos ||
url.find("api/v1/dbd-character-data/bulk-spending-bloodweb") != std::string::npos)
return onBloodweb(body, respHeaders);
}
void Spoofer::clientRequestHandler(std::string& url, const std::string& body, std::string& /*reqHeaders*/)
{
if (url.find("bhvrdbd.com") == std::string::npos) return;
std::lock_guard<std::mutex> lock(_mtx);
if (url.find("api/v1/dbd-character-data/bloodweb") != std::string::npos || if (url.find("api/v1/dbd-character-data/bloodweb") != std::string::npos ||
url.find("api/v1/dbd-character-data/bulk-spending-bloodweb") != std::string::npos) url.find("api/v1/dbd-character-data/bulk-spending-bloodweb") != std::string::npos)
{ {
try json req = json::parse(body, nullptr, false);
{ if (req.is_discarded()) return Log::error("JSON parse error for bloodweb request handler");
json req = json::parse(body); if (req.contains("characterName")) this->_lastBloodWebChar = req["characterName"];
if (req.contains("characterName"))
{
this->_lastBloodWebChar = req["characterName"];
Log::info("Detected bloodweb request for character: {}", this->_lastBloodWebChar);
}
}
catch (...)
{
}
} }
} }
+19 -5
View File
@@ -5,6 +5,8 @@
#include <unordered_set> #include <unordered_set>
#include <string> #include <string>
#include <mutex>
#include <nlohmann/json_fwd.hpp> #include <nlohmann/json_fwd.hpp>
class Spoofer class Spoofer
@@ -16,13 +18,22 @@ class Spoofer
void registerListeners(Proxy* proxy); void registerListeners(Proxy* proxy);
void loadData(); void loadData();
void parseCatalog(std::string data); bool parseCatalog(std::string data);
bool parseStackable(std::string data, std::unordered_set<std::string>& camperSet,
std::unordered_set<std::string>& slasherSet);
void parseAndDumpCatalog(std::string& data); std::string getRandomItem();
int getRandomQuantity();
void generateBloodweb(nlohmann::json& data);
void modifyCharacterData(nlohmann::json& js); void modifyCharacterData(nlohmann::json& js);
void onGetAll(std::string& body);
void onInventoryAll(std::string& body);
void onBloodweb(std::string& body, std::string& respHeaders);
void serverResponseHandler(const std::string& url, std::string& body, std::string& respHeaders); void serverResponseHandler(const std::string& url, std::string& body, std::string& respHeaders);
void clientRequestHandler(const std::string& url, const std::string& body, std::string& reqHeaders); void clientRequestHandler(std::string& url, const std::string& body, std::string& reqHeaders);
std::unordered_set<std::string> _camperItemIds; std::unordered_set<std::string> _camperItemIds;
std::unordered_set<std::string> _slasherPowerIds; std::unordered_set<std::string> _slasherPowerIds;
@@ -33,11 +44,14 @@ class Spoofer
std::unordered_set<std::string> _camperAddonIds; std::unordered_set<std::string> _camperAddonIds;
std::unordered_set<std::string> _slasherAddonIds; std::unordered_set<std::string> _slasherAddonIds;
std::unordered_set<std::string> _slasherPerkIds;
std::unordered_set<std::string> _camperPerkIds; std::unordered_set<std::string> _camperPerkIds;
std::unordered_set<std::string> _slasherPerkIds;
std::unordered_set<std::string> _catalogOutfitIds; std::unordered_set<std::string> _catalogOutfitIds;
std::unordered_set<std::string> _catalogItemIds; std::unordered_set<std::string> _catalogItemIds;
std::string _lastBloodWebChar = "Ace"; std::unordered_set<std::string> _unownedCharacters;
std::string _lastBloodWebChar = "";
std::mutex _mtx;
}; };
+14
View File
@@ -0,0 +1,14 @@
#include "utils.h"
#include <minwindef.h>
#include <libloaderapi.h>
std::string utils::getExePath()
{
char buffer[MAX_PATH];
GetModuleFileNameA(NULL, buffer, MAX_PATH);
std::string path(buffer);
size_t pos = path.find_last_of("\\/");
if (pos != std::string::npos) return path.substr(0, pos + 1);
return "";
}
+8
View File
@@ -0,0 +1,8 @@
#pragma once
#include <string>
namespace utils
{
std::string getExePath();
}