inital commit - WIP code

This commit is contained in:
Neo 2023-07-01 05:51:46 -07:00
parent a71c6c0e32
commit 37b23e0edc
12 changed files with 1570 additions and 0 deletions

926
.gitignore vendored Normal file
View File

@ -0,0 +1,926 @@
xpkgdb/
### venv template
# Virtualenv
# http://iamzed.com/2009/05/07/a-primer-on-virtualenv/
.Python
[Bb]in
[Ii]nclude
[Ll]ib
[Ll]ib64
[Ll]ocal
[Ss]cripts
pyvenv.cfg
.venv
pip-selfcheck.json
### NotepadPP template
# Notepad++ backups #
*.bak
### VisualStudioCode template
.vscode/*
!.vscode/settings.json
!.vscode/tasks.json
!.vscode/launch.json
!.vscode/extensions.json
!.vscode/*.code-snippets
# Local History for Visual Studio Code
.history/
# Built Visual Studio Code Extensions
*.vsix
### Linux template
*~
# temporary files which can be created if a process still has a handle open of a deleted file
.fuse_hidden*
# KDE directory preferences
.directory
# Linux trash folder which might appear on any partition or disk
.Trash-*
# .nfs files are created when an open file is removed but is still being accessed
.nfs*
### VisualStudio template
## Ignore Visual Studio temporary files, build results, and
## files generated by popular Visual Studio add-ons.
##
## Get latest from https://github.com/github/gitignore/blob/main/VisualStudio.gitignore
# User-specific files
*.rsuser
*.suo
*.user
*.userosscache
*.sln.docstates
# User-specific files (MonoDevelop/Xamarin Studio)
*.userprefs
# Mono auto generated files
mono_crash.*
# Build results
[Dd]ebug/
[Dd]ebugPublic/
[Rr]elease/
[Rr]eleases/
x64/
x86/
[Ww][Ii][Nn]32/
[Aa][Rr][Mm]/
[Aa][Rr][Mm]64/
bld/
[Bb]in/
[Oo]bj/
[Ll]og/
[Ll]ogs/
# Visual Studio 2015/2017 cache/options directory
.vs/
# Uncomment if you have tasks that create the project's static files in wwwroot
#wwwroot/
# Visual Studio 2017 auto generated files
Generated\ Files/
# MSTest test Results
[Tt]est[Rr]esult*/
[Bb]uild[Ll]og.*
# NUnit
*.VisualState.xml
TestResult.xml
nunit-*.xml
# Build Results of an ATL Project
[Dd]ebugPS/
[Rr]eleasePS/
dlldata.c
# Benchmark Results
BenchmarkDotNet.Artifacts/
# .NET Core
project.lock.json
project.fragment.lock.json
artifacts/
# ASP.NET Scaffolding
ScaffoldingReadMe.txt
# StyleCop
StyleCopReport.xml
# Files built by Visual Studio
*_i.c
*_p.c
*_h.h
*.ilk
*.meta
*.obj
*.iobj
*.pch
*.pdb
*.ipdb
*.pgc
*.pgd
*.rsp
*.sbr
*.tlb
*.tli
*.tlh
*.tmp
*.tmp_proj
*_wpftmp.csproj
*.log
*.tlog
*.vspscc
*.vssscc
.builds
*.pidb
*.svclog
*.scc
# Chutzpah Test files
_Chutzpah*
# Visual C++ cache files
ipch/
*.aps
*.ncb
*.opendb
*.opensdf
*.sdf
*.cachefile
*.VC.db
*.VC.VC.opendb
# Visual Studio profiler
*.psess
*.vsp
*.vspx
*.sap
# Visual Studio Trace Files
*.e2e
# TFS 2012 Local Workspace
$tf/
# Guidance Automation Toolkit
*.gpState
# ReSharper is a .NET coding add-in
_ReSharper*/
*.[Rr]e[Ss]harper
*.DotSettings.user
# TeamCity is a build add-in
_TeamCity*
# DotCover is a Code Coverage Tool
*.dotCover
# AxoCover is a Code Coverage Tool
.axoCover/*
!.axoCover/settings.json
# Coverlet is a free, cross platform Code Coverage Tool
coverage*.json
coverage*.xml
coverage*.info
# Visual Studio code coverage results
*.coverage
*.coveragexml
# NCrunch
_NCrunch_*
.*crunch*.local.xml
nCrunchTemp_*
# MightyMoose
*.mm.*
AutoTest.Net/
# Web workbench (sass)
.sass-cache/
# Installshield output folder
[Ee]xpress/
# DocProject is a documentation generator add-in
DocProject/buildhelp/
DocProject/Help/*.HxT
DocProject/Help/*.HxC
DocProject/Help/*.hhc
DocProject/Help/*.hhk
DocProject/Help/*.hhp
DocProject/Help/Html2
DocProject/Help/html
# Click-Once directory
publish/
# Publish Web Output
*.[Pp]ublish.xml
*.azurePubxml
# Note: Comment the next line if you want to checkin your web deploy settings,
# but database connection strings (with potential passwords) will be unencrypted
*.pubxml
*.publishproj
# Microsoft Azure Web App publish settings. Comment the next line if you want to
# checkin your Azure Web App publish settings, but sensitive information contained
# in these scripts will be unencrypted
PublishScripts/
# NuGet Packages
*.nupkg
# NuGet Symbol Packages
*.snupkg
# The packages folder can be ignored because of Package Restore
**/[Pp]ackages/*
# except build/, which is used as an MSBuild target.
!**/[Pp]ackages/build/
# Uncomment if necessary however generally it will be regenerated when needed
#!**/[Pp]ackages/repositories.config
# NuGet v3's project.json files produces more ignorable files
*.nuget.props
*.nuget.targets
# Microsoft Azure Build Output
csx/
*.build.csdef
# Microsoft Azure Emulator
ecf/
rcf/
# Windows Store app package directories and files
AppPackages/
BundleArtifacts/
Package.StoreAssociation.xml
_pkginfo.txt
*.appx
*.appxbundle
*.appxupload
# Visual Studio cache files
# files ending in .cache can be ignored
*.[Cc]ache
# but keep track of directories ending in .cache
!?*.[Cc]ache/
# Others
ClientBin/
~$*
*.dbmdl
*.dbproj.schemaview
*.jfm
*.pfx
*.publishsettings
orleans.codegen.cs
# Including strong name files can present a security risk
# (https://github.com/github/gitignore/pull/2483#issue-259490424)
#*.snk
# Since there are multiple workflows, uncomment next line to ignore bower_components
# (https://github.com/github/gitignore/pull/1529#issuecomment-104372622)
#bower_components/
# RIA/Silverlight projects
Generated_Code/
# Backup & report files from converting an old project file
# to a newer Visual Studio version. Backup files are not needed,
# because we have git ;-)
_UpgradeReport_Files/
Backup*/
UpgradeLog*.XML
UpgradeLog*.htm
ServiceFabricBackup/
*.rptproj.bak
# SQL Server files
*.mdf
*.ldf
*.ndf
# Business Intelligence projects
*.rdl.data
*.bim.layout
*.bim_*.settings
*.rptproj.rsuser
*- [Bb]ackup.rdl
*- [Bb]ackup ([0-9]).rdl
*- [Bb]ackup ([0-9][0-9]).rdl
# Microsoft Fakes
FakesAssemblies/
# GhostDoc plugin setting file
*.GhostDoc.xml
# Node.js Tools for Visual Studio
.ntvs_analysis.dat
node_modules/
# Visual Studio 6 build log
*.plg
# Visual Studio 6 workspace options file
*.opt
# Visual Studio 6 auto-generated workspace file (contains which files were open etc.)
*.vbw
# Visual Studio 6 auto-generated project file (contains which files were open etc.)
*.vbp
# Visual Studio 6 workspace and project file (working project files containing files to include in project)
*.dsw
*.dsp
# Visual Studio 6 technical files
# Visual Studio LightSwitch build output
**/*.HTMLClient/GeneratedArtifacts
**/*.DesktopClient/GeneratedArtifacts
**/*.DesktopClient/ModelManifest.xml
**/*.Server/GeneratedArtifacts
**/*.Server/ModelManifest.xml
_Pvt_Extensions
# Paket dependency manager
.paket/paket.exe
paket-files/
# FAKE - F# Make
.fake/
# CodeRush personal settings
.cr/personal
# Python Tools for Visual Studio (PTVS)
__pycache__/
*.pyc
# Cake - Uncomment if you are using it
# tools/**
# !tools/packages.config
# Tabs Studio
*.tss
# Telerik's JustMock configuration file
*.jmconfig
# BizTalk build output
*.btp.cs
*.btm.cs
*.odx.cs
*.xsd.cs
# OpenCover UI analysis results
OpenCover/
# Azure Stream Analytics local run output
ASALocalRun/
# MSBuild Binary and Structured Log
*.binlog
# NVidia Nsight GPU debugger configuration file
*.nvuser
# MFractors (Xamarin productivity tool) working folder
.mfractor/
# Local History for Visual Studio
.localhistory/
# Visual Studio History (VSHistory) files
.vshistory/
# BeatPulse healthcheck temp database
healthchecksdb
# Backup folder for Package Reference Convert tool in Visual Studio 2017
MigrationBackup/
# Ionide (cross platform F# VS Code tools) working folder
.ionide/
# Fody - auto-generated XML schema
FodyWeavers.xsd
# VS Code files for those working on multiple tools
*.code-workspace
# Local History for Visual Studio Code
# Windows Installer files from build outputs
*.cab
*.msi
*.msix
*.msm
*.msp
# JetBrains Rider
*.sln.iml
### Windows template
# Windows thumbnail cache files
Thumbs.db
Thumbs.db:encryptable
ehthumbs.db
ehthumbs_vista.db
# Dump file
*.stackdump
# Folder config file
[Dd]esktop.ini
# Recycle Bin used on file shares
$RECYCLE.BIN/
# Windows Installer files
# Windows shortcuts
*.lnk
### vs template
## Ignore Visual Studio temporary files, build results, and
## files generated by popular Visual Studio add-ons.
##
## Get latest from https://github.com/github/gitignore/blob/master/VisualStudio.gitignore
# User-specific files
# User-specific files (MonoDevelop/Xamarin Studio)
# Mono auto generated files
# Build results
# Visual Studio 2015/2017 cache/options directory
# Uncomment if you have tasks that create the project's static files in wwwroot
#wwwroot/
# Visual Studio 2017 auto generated files
# MSTest test Results
# NUnit
# Build Results of an ATL Project
# Benchmark Results
# .NET Core
# StyleCop
# Files built by Visual Studio
# Chutzpah Test files
# Visual C++ cache files
# Visual Studio profiler
# Visual Studio Trace Files
# TFS 2012 Local Workspace
# Guidance Automation Toolkit
# ReSharper is a .NET coding add-in
# TeamCity is a build add-in
# DotCover is a Code Coverage Tool
# AxoCover is a Code Coverage Tool
# Coverlet is a free, cross platform Code Coverage Tool
coverage*[.json, .xml, .info]
# Visual Studio code coverage results
# NCrunch
# MightyMoose
# Web workbench (sass)
# Installshield output folder
# DocProject is a documentation generator add-in
# Click-Once directory
# Publish Web Output
# Note: Comment the next line if you want to checkin your web deploy settings,
# but database connection strings (with potential passwords) will be unencrypted
# Microsoft Azure Web App publish settings. Comment the next line if you want to
# checkin your Azure Web App publish settings, but sensitive information contained
# in these scripts will be unencrypted
# NuGet Packages
# NuGet Symbol Packages
# The packages folder can be ignored because of Package Restore
# except build/, which is used as an MSBuild target.
# Uncomment if necessary however generally it will be regenerated when needed
#!**/[Pp]ackages/repositories.config
# NuGet v3's project.json files produces more ignorable files
# Microsoft Azure Build Output
# Microsoft Azure Emulator
# Windows Store app package directories and files
# Visual Studio cache files
# files ending in .cache can be ignored
# but keep track of directories ending in .cache
# Others
# Including strong name files can present a security risk
# (https://github.com/github/gitignore/pull/2483#issue-259490424)
#*.snk
# Since there are multiple workflows, uncomment next line to ignore bower_components
# (https://github.com/github/gitignore/pull/1529#issuecomment-104372622)
#bower_components/
# RIA/Silverlight projects
# Backup & report files from converting an old project file
# to a newer Visual Studio version. Backup files are not needed,
# because we have git ;-)
# SQL Server files
# Business Intelligence projects
# Microsoft Fakes
# GhostDoc plugin setting file
# Node.js Tools for Visual Studio
# Visual Studio 6 build log
# Visual Studio 6 workspace options file
# Visual Studio 6 auto-generated workspace file (contains which files were open etc.)
# Visual Studio LightSwitch build output
# Paket dependency manager
# FAKE - F# Make
# CodeRush personal settings
# Python Tools for Visual Studio (PTVS)
# Cake - Uncomment if you are using it
# tools/**
# !tools/packages.config
# Tabs Studio
# Telerik's JustMock configuration file
# BizTalk build output
# OpenCover UI analysis results
# Azure Stream Analytics local run output
# MSBuild Binary and Structured Log
# NVidia Nsight GPU debugger configuration file
# MFractors (Xamarin productivity tool) working folder
# Local History for Visual Studio
# BeatPulse healthcheck temp database
# Backup folder for Package Reference Convert tool in Visual Studio 2017
# Ionide (cross platform F# VS Code tools) working folder
### macOS template
# General
.DS_Store
.AppleDouble
.LSOverride
# Icon must end with two \r
Icon
# Thumbnails
._*
# Files that might appear in the root of a volume
.DocumentRevisions-V100
.fseventsd
.Spotlight-V100
.TemporaryItems
.Trashes
.VolumeIcon.icns
.com.apple.timemachine.donotpresent
# Directories potentially created on remote AFP share
.AppleDB
.AppleDesktop
Network Trash Folder
Temporary Items
.apdisk
### PyCharm+all template
# Covers JetBrains IDEs: IntelliJ, RubyMine, PhpStorm, AppCode, PyCharm, CLion, Android Studio, WebStorm and Rider
# Reference: https://intellij-support.jetbrains.com/hc/en-us/articles/206544839
# User-specific stuff
.idea/**/workspace.xml
.idea/**/tasks.xml
.idea/**/usage.statistics.xml
.idea/**/dictionaries
.idea/**/shelf
# AWS User-specific
.idea/**/aws.xml
# Generated files
.idea/**/contentModel.xml
# Sensitive or high-churn files
.idea/**/dataSources/
.idea/**/dataSources.ids
.idea/**/dataSources.local.xml
.idea/**/sqlDataSources.xml
.idea/**/dynamic.xml
.idea/**/uiDesigner.xml
.idea/**/dbnavigator.xml
# Gradle
.idea/**/gradle.xml
.idea/**/libraries
# Gradle and Maven with auto-import
# When using Gradle or Maven with auto-import, you should exclude module files,
# since they will be recreated, and may cause churn. Uncomment if using
# auto-import.
# .idea/artifacts
# .idea/compiler.xml
# .idea/jarRepositories.xml
# .idea/modules.xml
# .idea/*.iml
# .idea/modules
# *.iml
# *.ipr
# CMake
cmake-build-*/
# Mongo Explorer plugin
.idea/**/mongoSettings.xml
# File-based project format
*.iws
# IntelliJ
out/
# mpeltonen/sbt-idea plugin
.idea_modules/
# JIRA plugin
atlassian-ide-plugin.xml
# Cursive Clojure plugin
.idea/replstate.xml
# SonarLint plugin
.idea/sonarlint/
# Crashlytics plugin (for Android Studio and IntelliJ)
com_crashlytics_export_strings.xml
crashlytics.properties
crashlytics-build.properties
fabric.properties
# Editor-based Rest Client
.idea/httpRequests
# Android studio 3.1+ serialized cache file
.idea/caches/build_file_checksums.ser
### Python template
# Byte-compiled / optimized / DLL files
*.py[cod]
*$py.class
# C extensions
*.so
# Distribution / packaging
build/
develop-eggs/
dist/
downloads/
eggs/
.eggs/
lib/
lib64/
parts/
sdist/
var/
wheels/
share/python-wheels/
*.egg-info/
.installed.cfg
*.egg
MANIFEST
# PyInstaller
# Usually these files are written by a python script from a template
# before PyInstaller builds the exe, so as to inject date/other infos into it.
*.manifest
*.spec
# Installer logs
pip-log.txt
pip-delete-this-directory.txt
# Unit test / coverage reports
htmlcov/
.tox/
.nox/
.coverage
.coverage.*
.cache
nosetests.xml
coverage.xml
*.cover
*.py,cover
.hypothesis/
.pytest_cache/
cover/
# Translations
*.mo
*.pot
# Django stuff:
local_settings.py
db.sqlite3
db.sqlite3-journal
# Flask stuff:
instance/
.webassets-cache
# Scrapy stuff:
.scrapy
# Sphinx documentation
docs/_build/
# PyBuilder
.pybuilder/
target/
# Jupyter Notebook
.ipynb_checkpoints
# IPython
profile_default/
ipython_config.py
# pyenv
# For a library or package, you might want to ignore these files since the code is
# intended to run in multiple environments; otherwise, check them in:
# .python-version
# pipenv
# According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control.
# However, in case of collaboration, if having platform-specific dependencies or dependencies
# having no cross-platform support, pipenv may install dependencies that don't work, or not
# install all needed dependencies.
#Pipfile.lock
# poetry
# Similar to Pipfile.lock, it is generally recommended to include poetry.lock in version control.
# This is especially recommended for binary packages to ensure reproducibility, and is more
# commonly ignored for libraries.
# https://python-poetry.org/docs/basic-usage/#commit-your-poetrylock-file-to-version-control
#poetry.lock
# pdm
# Similar to Pipfile.lock, it is generally recommended to include pdm.lock in version control.
#pdm.lock
# pdm stores project-wide configurations in .pdm.toml, but it is recommended to not include it
# in version control.
# https://pdm.fming.dev/#use-with-ide
.pdm.toml
# PEP 582; used by e.g. github.com/David-OConnor/pyflow and github.com/pdm-project/pdm
__pypackages__/
# Celery stuff
celerybeat-schedule
celerybeat.pid
# SageMath parsed files
*.sage.py
# Environments
.env
env/
venv/
ENV/
env.bak/
venv.bak/
# Spyder project settings
.spyderproject
.spyproject
# Rope project settings
.ropeproject
# mkdocs documentation
/site
# mypy
.mypy_cache/
.dmypy.json
dmypy.json
# Pyre type checker
.pyre/
# pytype static type analyzer
.pytype/
# Cython debug symbols
cython_debug/
# PyCharm
# JetBrains specific template is maintained in a separate JetBrains.gitignore that can
# be found at https://github.com/github/gitignore/blob/main/Global/JetBrains.gitignore
# and can be added to the global gitignore or merged into this file. For a more nuclear
# option (not recommended) you can uncomment the following to ignore the entire idea folder.
#.idea/
### SublimeText template
# Cache files for Sublime Text
*.tmlanguage.cache
*.tmPreferences.cache
*.stTheme.cache
# Workspace files are user-specific
*.sublime-workspace
# Project files should be checked into the repository, unless a significant
# proportion of contributors will probably not be using Sublime Text
# *.sublime-project
# SFTP configuration file
sftp-config.json
sftp-config-alt*.json
# Package control specific files
Package Control.last-run
Package Control.ca-list
Package Control.ca-bundle
Package Control.system-ca-bundle
Package Control.cache/
Package Control.ca-certs/
Package Control.merged-ca-bundle
Package Control.user-ca-bundle
oscrypto-ca-bundle.crt
bh_unicode_properties.cache
# Sublime-github package stores a github token in this file
# https://packagecontrol.io/packages/sublime-github
GitHub.sublime-settings

View File

@ -1,2 +1,9 @@
# DB
Scripts used to make the UMSKT keys.json DB
## Requirements
* python
* sagemath
* pefile

110
jsonExport Normal file
View File

@ -0,0 +1,110 @@
#!/usr/bin/env python
import sqlite3
import json
def serialize_binary_data(data):
if isinstance(data, bytes):
return str(data)
elif isinstance(data, list):
return [serialize_binary_data(item) for item in data]
elif isinstance(data, dict):
return {key: serialize_binary_data(value) for key, value in data.items()}
return data
def dump_sqlite_to_json(database_file, output_file):
connection = sqlite3.connect(database_file)
connection.row_factory = sqlite3.Row
cursor = connection.cursor()
tables = [
"BINK",
"DPCDLL",
"DPCDLL_ENTRIES",
"LICDLL",
"PIDGEN",
"BINK_PIDGEN",
"PRODUCT"
]
data = {}
for table in tables:
cursor.execute(f"SELECT * FROM {table}")
rows = cursor.fetchall()
if table == "BINK_PIDGEN":
merged_data = {}
for row in rows:
pidgen_id = row["PIDGEN"]
bink = row["BINK"]
if pidgen_id in merged_data:
merged_data[pidgen_id]["BINK"].append(bink)
else:
merged_data[pidgen_id] = {}
merged_data[pidgen_id]["BINK"] = [bink]
serialized_merged_data = serialize_binary_data(merged_data)
data[table] = serialized_merged_data
elif table == "PRODUCT":
product_data = {}
for row in rows:
shortname = row["shortname"]
product_data[shortname] = serialize_binary_data(dict(row))
data[table] = product_data
elif table == "LICDLL":
licdll_data = {}
for row in rows:
row_dict = dict(row)
id_value = row_dict.pop("ID", None)
licdll_data[id_value] = serialize_binary_data(row_dict)
data[table] = licdll_data
elif table == "DPCDLL_ENTRIES":
dpcll_entries_data = {}
for row in rows:
dpcll_id = row["DPCDLL_ID"]
bink_id = row["BINK_ID"]
entry_data = dict(row)
del entry_data["DPCDLL_ID"]
del entry_data["BINK_ID"]
if dpcll_id not in dpcll_entries_data:
dpcll_entries_data[dpcll_id] = {}
dpcll_entries_data[dpcll_id][bink_id] = serialize_binary_data(entry_data)
data[table] = dpcll_entries_data
elif table == "DPCDLL":
dpcll_data = {}
for row in rows:
row_dict = dict(row)
id_value = row_dict.pop("ID", None)
dpcll_data[id_value] = serialize_binary_data(row_dict)
data[table] = dpcll_data
elif table == "BINK":
bink_data = {}
for row in rows:
row_dict = dict(row)
id_value = row_dict.pop("ID", None)
bink_data[id_value] = serialize_binary_data(row_dict)
data[table] = bink_data
else:
serialized_rows = [serialize_binary_data(dict(row)) for row in rows]
data[table] = serialized_rows
with open(output_file, 'w') as file:
json.dump(data, file, indent=4)
cursor.close()
connection.close()
# Usage example
database_file = "xpkgdb.sqlite"
output_file = "xpkgdb.json"
dump_sqlite_to_json(database_file, output_file)

226
main.py Normal file
View File

@ -0,0 +1,226 @@
#!/usr/bin/env sage
import os
import sys
import sqlite3
import parser
import tempfile
import re
import subprocess
import hashlib
import traceback
def extract_file_with_7z(contianer_path, file_path):
try:
output = subprocess.check_output(['7z', 'x', '-so', contianer_path, file_path], stderr=subprocess.STDOUT)
except subprocess.CalledProcessError as e:
output = e.output
return output
def process_file_contents_with_7z(container_path, file_path, process_func, container_extension=None):
process_func(extract_file_with_7z(container_path, file_path), file_path, container_extension)
def process_container_with_7z(file_path):
files = {}
# Use 7zip to list the contents of the MSI file
command = ['7z', 'l', '-slt', '-r', '-sdel', '-so', file_path]
try:
output = subprocess.check_output(command, universal_newlines=True, stderr=subprocess.STDOUT)
except subprocess.CalledProcessError as e:
output = e.output
# Parse the output
lines = output.splitlines()
i = 0
while i < len(lines):
file = {'path': '', 'size': 0, 'packed_size': 0, 'created': '', 'modified': ''}
if lines[i].startswith('Path ='):
# Extract the path, size, packed size, created, and modified information
file['path'] = re.search(r'Path = (.+)', lines[i]).group(1)
while i < len(lines):
line = lines[i]
try:
if line.startswith('Size = '):
file['size'] = re.search(r'Size = (\d+)', line).group(1)
elif line.startswith('Packed Size = '):
file['packed_size'] = re.search(r'Packed Size = (\d+)', line).group(1)
elif line.startswith('Created = '):
file['created'] = re.search(r'Created = (.+)', line).group(1)
elif line.startswith('Modified = '):
file['modified'] = re.search(r'Modified = (.+)', line).group(1)
elif line == "":
break
except AttributeError:
i = i
i += 1
files[file['path']] = file
i += 1
return files
file_prefixes = ['pidgen', 'licdll', 'dpcdll', 'mso', 'msa', 'pidca']
def process_nested_file(temp_container_path, path):
path_lower = path.lower()
if any(path_lower.startswith(prefix) for prefix in file_prefixes):
if path_lower.endswith('dll'):
compressed_file_data = extract_file_with_7z(temp_container_path, path)
process_dll(compressed_file_data, path)
if path_lower.endswith('dl_'):
compressed_file_data = extract_file_with_7z(temp_container_path, path)
process_container(compressed_file_data, path, container_extension='.dl_')
def process_container(file_data, file_path, container_extension=None):
# Create a temporary file
with tempfile.NamedTemporaryFile(suffix=container_extension, delete=False) as temp_container_file:
temp_container_path = temp_container_file.name
temp_container_file.write(file_data)
temp_container_file.close()
files = process_container_with_7z(temp_container_path)
if container_extension == '.msi':
for path, file in files.items():
process_nested_file(temp_container_path, path)
if path.lower().startswith('binary.'):
# Read the contents of files starting with 'Binary.'
print(f'Parsing MSI Stream Name: {path}')
process_file_contents_with_7z(temp_container_path, path, process_dll)
if container_extension == '.cab':
for path, file in files.items():
process_nested_file(temp_container_path, path)
if container_extension == '.dl_':
process_file_contents_with_7z(temp_container_path, file_path, process_dll)
# Remove the temporary container file
os.remove(temp_container_path)
def process_dll(file_data, file_path, container_extension=None):
# Process the DLL file as needed
print(f'[{file_path}]: Parsing file')
pidgen_data = parser.pidgen.parse(file_data)
if pidgen_data != {}:
print(f'[{file_path}]: Found PIDGEN data')
sha1 = hashlib.sha1(file_data).hexdigest()
print(f'[{file_path}]: SHA1: {sha1}')
print(pidgen_data)
try:
dpcll_data = parser.dpcdll.parse(file_data)
if dpcll_data != {}:
print(f'[{file_path}]: Found DPCDLL data')
sha1 = hashlib.sha1(file_data).hexdigest()
print(f'[{file_path}]: SHA1: {sha1}')
except ValueError:
dpcll_data = {}
if any(file_path.lower().startswith(prefix) for prefix in ['licdll', 'mso.dll', 'msa.dll']):
print(f'[{file_path}]: Cataloguing a LICDLL type file')
sha1 = hashlib.sha1(file_data).hexdigest()
print(f'[{file_path}]: SHA1: {sha1}')
def process_iso(file_path):
files = process_container_with_7z(file_path)
for path, file in files.items():
if path.lower().endswith('.msi'):
print(f'[{path}]: Processing MSI file')
process_file_contents_with_7z(file_path, path, process_container, container_extension='.msi')
if path.lower().endswith('.cab'):
print(f'[{path}]: Processing CAB file')
process_file_contents_with_7z(file_path, path, process_container, container_extension='.cab')
if path.lower().endswith('.dll'):
print(f'[{path}]: Processing DLL file')
process_file_contents_with_7z(file_path, path, process_dll)
def process_file_or_folder(path):
extensions = ['.iso', '.img']
if os.path.isfile(path):
print(f'[{path}]: Processing ISO/Disk Image file')
process_iso(path)
elif os.path.isdir(path):
print(f'[{path}]: Recursing through folder')
for root, dirs, files in os.walk(path):
for file in files:
if file.lower().endswith('.iso'):
iso_path = os.path.join(root, file)
print(f'Processing ISO file: {iso_path}')
process_iso(iso_path)
else:
print(f'Invalid file or folder: {path}')
def check_7z_command():
if sys.platform.startswith('win'): # Windows
try:
# Use the 'where' command to check if '7z' is in the path
subprocess.check_output(['where', '7z'])
return True
except subprocess.CalledProcessError:
return False
else: # Unix-based systems (Linux, macOS, etc.)
try:
# Use the 'which' command to check if '7z' is in the path
subprocess.check_output(['which', '7z'])
return True
except subprocess.CalledProcessError:
return False
# Main function
def main():
if len(sys.argv) != 3:
print('Usage: {} <file.iso|folder> <database>'.format(sys.argv[0]))
print('Parses <file.iso|folder> for various DLLs required for product licensing and activation')
print('Data is saved to the SQLite3 Database <database> and will be created if it does not exist')
print('If a <folder> is specified, it will search recursively for files ending in .iso/.img')
sys.exit(1)
path = sys.argv[1]
database = sys.argv[2]
if not os.path.exists(database):
conn = sqlite3.connect(database)
with open('newdb.sql') as f:
conn.executescript(f.read())
conn.close()
conn = sqlite3.connect(database)
process_file_or_folder(path)
conn.close()
# Entry point
if __name__ == '__main__':
if not check_7z_command():
print('7zip is not in the path, please add the 7z executable to the system path and try again')
try:
main()
except Exception as e:
print('An error occurred:', e)
traceback.print_exc()
sys.exit(1)

0
newdb.sql Normal file
View File

4
parser/__init__.py Normal file
View File

@ -0,0 +1,4 @@
import parser.dpcdll as dpcdll
import parser.pidgen as pidgen
__all__ = ['dpcdll', 'pidgen']

View File

@ -0,0 +1,3 @@
from .dpcdll import parse
__all__ = ['parse']

61
parser/dpcdll/dpcdll.py Normal file
View File

@ -0,0 +1,61 @@
from struct import unpack
def read_int(data, offset):
return unpack('<I', data[offset:offset + 4])[0]
def parse(file_data):
lic_types = ['NULL', 'Volume', 'Retail', 'Evaluation', 'Tablet', 'OEM', 'Embedded']
dpc_data = {}
tmp = file_data
offset = tmp.find(b'\x00\x00\x00\xff\xff\xff\x7f\x80') - 21
del tmp
if offset == -22:
raise ValueError('Offset not found')
while file_data[offset:offset + 4] != b'\x00\x00\x00\x00':
offset -= 164
offset -= 4
while True:
if offset < 0 or offset + 32 >= len(file_data):
raise ValueError('Error in offset or not enough data')
ind = read_int(file_data, offset)
bink_id = hex(read_int(file_data, offset + 4)).zfill(4).upper()
min_pid = read_int(file_data, offset + 8)
max_pid = read_int(file_data, offset + 12)
if min_pid > 999 or max_pid > 999:
break
lic_type = read_int(file_data, offset + 16)
if lic_type > 6:
break
days_to_act = str(read_int(file_data, offset + 20))
eval_days = str(read_int(file_data, offset + 24))
sig_len = read_int(file_data, offset + 28)
if offset + 32 + sig_len >= len(file_data):
raise ValueError('Error in signature length or not enough data')
offset += 32 + sig_len
if bink_id not in dpc_data:
dpc_data[bink_id] = []
dpc_data[bink_id].append({
'Type': lic_types[lic_type],
'PIDRange': [min_pid, max_pid],
'EvaluationDays': eval_days,
'ActivationGraceDays': days_to_act
})
return dpc_data

View File

@ -0,0 +1,3 @@
from .pidgen import parse
__all__ = ['parse']

View File

@ -0,0 +1,4 @@
from .ecdlpsolver import parse
from .ecdlpsolver import parse_bink_data_with_sage
__all__ = ['parse', 'parse_bink_data_with_sage']

View File

@ -0,0 +1,131 @@
try:
from sage.all import *
except ImportError:
print("Unable to load SageMath")
# def warnx(*args, **kwargs):
# print(*args, file=sys.stderr, **kwargs)
# def tee(*output, file=None, **kwargs):
# print(*output, file=sys.stdout, **kwargs)
# if file is not None:
# print(*output, file=file, **kwargs)
def btoi(bb):
return int.from_bytes(bb, byteorder='little')
def rfactor(m, keysize, B):
digits = len('%d' % (2 ^ keysize - 1))
ff = ecm.find_factor(m, factor_digits=digits) # Try to find a good candidate
for f in ff:
if f > 2 and f.is_prime() and not f * B:
# warnx("ok for %d" % f)
return True, [f]
else:
# warnx("bad run: %s" % ff)
return False, ff
def parse_bink_data_with_sage(bink):
curve = bink['curve']
bink_header = bink['header']
F = GF(curve['p'])
# warnx("offs = %d, nb = %d, p = %x" % (offs, nb, p))
a = F(curve['a'])
b = F(curve['b'])
bx = F(curve['g']['x'])
by = F(curve['g']['y'])
Kx = F(curve['pub']['x'])
Ky = F(curve['pub']['y'])
E = EllipticCurve(F, [0, 0, 0, a, b])
# warnx(E)
B = E(bx, by)
K = E(Kx, Ky)
# If we get here, we know B and K are on the curve.
# Now get the order of the curve and then factorize it.
n = E.order()
# warnx("n = %d, now factoring..." % n)
# Find L by just trying if any of the factors in f yield the point at infinity
factors = []
ok, values = rfactor(n, bink_header['hashlen'], B)
while not ok:
for value in values:
ok, nl = rfactor(value, bink_header['keysize'], B)
if ok:
L = nl[0]
break
values.extend(nl)
factors = [n // L, L]
# warnx(factors)
# warnx("Reduce the result of ECDLP Solver modulo %d" % L)
# warnx("\n\njob input:\n\n")
bink['curve']['n'] = L
solver_input = ''
solver_input += 'GF := GF(%d);\n' % curve['p']
solver_input += 'E := EllipticCurve([GF|%d,%d]);\n' % (curve['a'], curve['b'])
solver_input += 'G := E![%d,%d];\n' % (curve['g']['x'], curve['g']['y'])
solver_input += 'K := E![%d,%d];\n' % (curve['pub']['x'], curve['pub']['y'])
solver_input += '/*\n'
solver_input += 'FactorCount:=%d;\n' % len(factors)
for f in factors:
solver_input += '%d;\n' % f
solver_input += '*/'
bink['solver_input'] = solver_input
return bink
def parse(bink):
bink_id_int = btoi(bink[0x00:0x04])
bink_id = "%08x" % bink_id_int
bink_header = {
'identifier': btoi(bink[0x00:0x04]),
'sizeof': btoi(bink[0x04:0x08]),
'countof': btoi(bink[0x08:0x0C]),
'checksum': btoi(bink[0x0C:0x10]),
'version': btoi(bink[0x10:0x14]),
'keysize': btoi(bink[0x14:0x18]),
'hashlen': btoi(bink[0x18:0x1C]),
'siglen': btoi(bink[0x1C:0x20]),
}
bink_values = bink[0x20:]
if bink_header["version"] > 20020420:
bink_values = bink[0x28:]
bink_header['authlen'] = btoi(bink[0x20:0x24])
bink_header['pidlen'] = btoi(bink[0x24:0x28])
offs = bink_header["keysize"] * 4
curve = {
'p': btoi(bink_values[:offs]),
'a': btoi(bink_values[offs:offs * 2]),
'b': btoi(bink_values[offs * 2:offs * 3]),
'g': {'x': btoi(bink_values[offs * 3:offs * 4]), 'y': btoi(bink_values[offs * 4:offs * 5])},
'pub': {'x': btoi(bink_values[offs * 5:offs * 6]), 'y': btoi(bink_values[offs * 6:offs * 7])},
'n': '',
'priv': ''
}
output = {
'bink_name': bink_id,
'header': bink_header,
'curve': curve,
}
output = parse_bink_data_with_sage(output)
return output

95
parser/pidgen/pidgen.py Normal file
View File

@ -0,0 +1,95 @@
import hashlib
import parser.pidgen.bink as bink
import pefile
def bink_out(resource_data):
# Process the resource data
# bink_id = "%02x" % int.from_bytes(resource_data[0x00:0x04], 'little')
sha1_hash = hashlib.sha1(resource_data).hexdigest()
data = bink.parse(resource_data)
data["sha1_hash"] = sha1_hash
return data
def parse(file_data):
output = {}
found_bink = 0
# attempt extracting via the PE directory
try:
# Load the DLL using pefile
pe = pefile.PE(data=file_data)
# Loop over the resources in the DLL
for resource_type in pe.DIRECTORY_ENTRY_RESOURCE.entries:
if resource_type.name is not None and resource_type.name.string.decode() == 'BINK':
found_bink = 1
# Extract resources from the "BINK" resource type
for resource_id_entry in resource_type.directory.entries:
for resource_entry in resource_id_entry.directory.entries:
resource_offset = resource_entry.data.struct.OffsetToData
resource_size = resource_entry.data.struct.Size
# Access the resource data
resource_data = pe.get_memory_mapped_image()[resource_offset: resource_offset + resource_size]
bink_id = "%08x" % int.from_bytes(resource_data[0x00:0x04], 'little')
output[bink_id] = bink_out(resource_data)
# Close the PE file
pe.close()
except pefile.PEFormatError as e:
found_bink = 0
except AttributeError as e:
found_bink = 0
# attempt a string search
if found_bink == 0:
string_1998 = b'\xAE\xDF\x30\x01'
string_2002 = b'\xC4\x7C\x31\x01'
entries = {}
for i in range(len(file_data) - 3):
if ((file_data[i:i + 4] == string_1998 and i + 0x170 < len(file_data) and (
file_data[i + 0x170:i + 0x170 + 4] == string_1998)) or (
file_data[i:i + 4] == string_1998 and i - 0x170 > 0 and (
file_data[i - 0x170:i - 0x170 + 4] == string_1998))):
start = i - 16
end = start + int.from_bytes(file_data[start + 4:start + 8], 'little') + 4
entries[start] = {
"Type": "BINK1998",
"StartAddress": start,
"EndAddress": end
}
if ((file_data[i:i + 4] == string_1998 and i + 0x180 < len(file_data) and (
file_data[i + 0x180:i + 0x180 + 4] == string_1998)) or (
file_data[i:i + 4] == string_1998 and i - 0x180 > 0 and (
file_data[i - 0x180:i - 0x180 + 4] == string_1998))):
start = i - 16
end = start + int.from_bytes(file_data[start + 4:start + 8], 'little') + 4
entries[start] = {
"Type": "BINK1998",
"StartAddress": start,
"EndAddress": end
}
elif file_data[i:i + 4] == string_2002 and i + 0x1E8 < len(file_data) and (
file_data[i + 0x1E8:i + 0x1E8 + 4] == string_2002):
start = i - 16
end = start + int.from_bytes(file_data[start + 4:start + 8], 'little') + 4
entries[start] = {
"Type": "BINK2002",
"StartAddress": start,
"EndAddress": end
}
if len(entries) != 0:
for key, value in entries.items():
bink_data = file_data[key:value['EndAddress']]
bink_id = "%08x" % int.from_bytes(bink_data[0x00:0x04], 'little')
output[bink_id] = bink_out(bink_data)
return output