mirror of
https://github.com/Sneed-Group/Poodletooth-iLand
synced 2025-01-09 17:53:50 +00:00
Cleanup
This commit is contained in:
parent
0628201fca
commit
0d198f9f5d
8 changed files with 107 additions and 129 deletions
|
@ -14,9 +14,9 @@ for line in prc.split('\n'):
|
|||
line = line.strip()
|
||||
if line:
|
||||
loadPrcFileData('nirai config', line)
|
||||
|
||||
|
||||
del prc
|
||||
|
||||
|
||||
# DC
|
||||
__builtin__.dcStream = StringStream()
|
||||
|
||||
|
@ -39,18 +39,18 @@ abort = False
|
|||
for mf in mfs:
|
||||
filename = 'resources/default/phase_%s.mf' % mf
|
||||
if not os.path.isfile(filename):
|
||||
print 'Phase %s not found' % filename
|
||||
print 'Phase %s not found' % filename
|
||||
abort = True
|
||||
break
|
||||
|
||||
|
||||
mf = Multifile()
|
||||
mf.openRead(filename)
|
||||
|
||||
|
||||
if not vfs.mount(mf, '../resources', 0):
|
||||
print 'Unable to mount %s' % filename
|
||||
abort = True
|
||||
break
|
||||
|
||||
|
||||
# Packs
|
||||
pack = os.environ.get('TT_STRIDE_CONTENT_PACK')
|
||||
if pack and pack != 'default':
|
||||
|
@ -63,14 +63,14 @@ if pack and pack != 'default':
|
|||
ext = os.path.splitext(name)[1]
|
||||
if ext not in ('.jpg', '.jpeg', '.ogg', '.rgb'):
|
||||
mf.removeSubfile(name)
|
||||
|
||||
|
||||
mf.flush()
|
||||
|
||||
|
||||
if not vfs.mount(mf, '../resources', 0):
|
||||
print 'Unable to mount %s' % filename
|
||||
abort = True
|
||||
break
|
||||
|
||||
|
||||
if not abort:
|
||||
# Run
|
||||
import toontown.toonbase.ClientStart
|
||||
|
|
|
@ -11,75 +11,75 @@ parser.add_argument('--make-nri', '-n', action='store_true',
|
|||
help='Generate stride NRI.')
|
||||
args = parser.parse_args()
|
||||
|
||||
# BEGIN (STRIPPED AND MODIFIED) COPY FROM niraitools.py
|
||||
# BEGIN (STRIPPED AND MODIFIED) COPY FROM niraitools.py
|
||||
class NiraiPackager:
|
||||
HEADER = 'NRI\n'
|
||||
|
||||
|
||||
def __init__(self, outfile):
|
||||
self.modules = {}
|
||||
self.outfile = outfile
|
||||
|
||||
|
||||
def __read_file(self, filename, mangler=None):
|
||||
with open(filename, 'rb') as f:
|
||||
data = f.read()
|
||||
|
||||
|
||||
base = filename.rsplit('.', 1)[0].replace('\\', '/').replace('/', '.')
|
||||
pkg = base.endswith('.__init__')
|
||||
moduleName = base.rsplit('.', 1)[0] if pkg else base
|
||||
|
||||
|
||||
name = moduleName
|
||||
if mangler is not None:
|
||||
name = mangler(name)
|
||||
|
||||
|
||||
if not name:
|
||||
return '', ('', 0)
|
||||
|
||||
|
||||
try:
|
||||
data = self.compile_module(name, data)
|
||||
|
||||
|
||||
except:
|
||||
print 'WARNING: Failed to compile', filename
|
||||
return '', ('', 0)
|
||||
|
||||
|
||||
size = len(data) * (-1 if pkg else 1)
|
||||
return name, (data, size)
|
||||
|
||||
|
||||
def compile_module(self, name, data):
|
||||
return marshal.dumps(compile(data, name, 'exec'))
|
||||
|
||||
|
||||
def add_module(self, moduleName, data, size=None, compile=False):
|
||||
if compile:
|
||||
data = self.compile_module(moduleName, data)
|
||||
|
||||
|
||||
if size is None:
|
||||
size = len(data)
|
||||
|
||||
|
||||
self.modules[moduleName] = (data, size)
|
||||
|
||||
|
||||
def add_file(self, filename, mangler=None):
|
||||
print 'Adding file', filename
|
||||
moduleName, (data, size) = self.__read_file(filename, mangler)
|
||||
if moduleName:
|
||||
moduleName = os.path.basename(filename).rsplit('.', 1)[0]
|
||||
self.add_module(moduleName, data, size)
|
||||
|
||||
|
||||
def add_directory(self, dir, mangler=None):
|
||||
print 'Adding directory', dir
|
||||
|
||||
|
||||
def _recurse_dir(dir):
|
||||
for f in os.listdir(dir):
|
||||
f = os.path.join(dir, f)
|
||||
|
||||
|
||||
if os.path.isdir(f):
|
||||
_recurse_dir(f)
|
||||
|
||||
|
||||
elif f.endswith('py'):
|
||||
moduleName, (data, size) = self.__read_file(f, mangler)
|
||||
if moduleName:
|
||||
self.add_module(moduleName, data, size)
|
||||
|
||||
|
||||
_recurse_dir(dir)
|
||||
|
||||
|
||||
def get_mangle_base(self, *path):
|
||||
return len(os.path.join(*path).rsplit('.', 1)[0].replace('\\', '/').replace('/', '.')) + 1
|
||||
|
||||
|
@ -88,112 +88,112 @@ class NiraiPackager:
|
|||
f.write(self.HEADER)
|
||||
f.write(self.process_modules())
|
||||
f.close()
|
||||
|
||||
|
||||
def generate_key(self, size=256):
|
||||
return os.urandom(size)
|
||||
|
||||
|
||||
def dump_key(self, key):
|
||||
for k in key:
|
||||
print ord(k),
|
||||
|
||||
|
||||
print
|
||||
|
||||
|
||||
def process_modules(self):
|
||||
# Pure virtual
|
||||
raise NotImplementedError('process_datagram')
|
||||
|
||||
|
||||
def get_file_contents(self, filename, keysize=0):
|
||||
with open(filename, 'rb') as f:
|
||||
data = f.read()
|
||||
|
||||
|
||||
if keysize:
|
||||
key = self.generate_key(keysize)
|
||||
rc4.rc4_setkey(key)
|
||||
data = key + rc4.rc4(data)
|
||||
|
||||
|
||||
return data
|
||||
# END COPY FROM niraitools.py
|
||||
# END COPY FROM niraitools.py
|
||||
|
||||
class StridePackager(NiraiPackager):
|
||||
HEADER = 'STRIDETT'
|
||||
BASEDIR = '..' + os.sep
|
||||
|
||||
|
||||
def __init__(self, outfile):
|
||||
NiraiPackager.__init__(self, outfile)
|
||||
self.__manglebase = self.get_mangle_base(self.BASEDIR)
|
||||
|
||||
|
||||
def add_source_dir(self, dir):
|
||||
self.add_directory(self.BASEDIR + dir, mangler=self.__mangler)
|
||||
|
||||
|
||||
def add_data_file(self, file):
|
||||
mb = self.get_mangle_base('data/')
|
||||
self.add_file('data/%s.py' % file, mangler=lambda x: x[mb:])
|
||||
|
||||
def __mangler(self, name):
|
||||
|
||||
def __mangler(self, name):
|
||||
if name.endswith('AI') or name.endswith('UD') or name in ('ToontownAIRepository', 'ToontownUberRepository',
|
||||
'ToontownInternalRepository'):
|
||||
if not 'NonRepeatableRandomSource' in name:
|
||||
return ''
|
||||
|
||||
return name[self.__manglebase:].strip('.')
|
||||
|
||||
|
||||
def generate_niraidata(self):
|
||||
print 'Generating niraidata'
|
||||
|
||||
|
||||
config = self.get_file_contents('../dependencies/config/release/en.prc')
|
||||
config += '\n\n' + self.get_file_contents('../dependencies/config/general.prc')
|
||||
key = self.generate_key(128)
|
||||
rc4.rc4_setkey(key)
|
||||
config = key + rc4.rc4(config)
|
||||
|
||||
|
||||
niraidata = 'CONFIG = %r' % config
|
||||
niraidata += '\nDC = %r' % self.get_file_contents('../dependencies/astron/dclass/stride.dc', 128)
|
||||
self.add_module('niraidata', niraidata, compile=True)
|
||||
|
||||
|
||||
def process_modules(self):
|
||||
with open('base.dg', 'rb') as f:
|
||||
basesize, = struct.unpack('<I', f.read(4))
|
||||
data = f.read()
|
||||
|
||||
|
||||
dg = Datagram()
|
||||
dg.addUint32(len(self.modules) + basesize)
|
||||
dg.appendData(data)
|
||||
|
||||
|
||||
for moduleName in self.modules:
|
||||
data, size = self.modules[moduleName]
|
||||
|
||||
|
||||
dg.addString(moduleName)
|
||||
dg.addInt32(size)
|
||||
dg.appendData(data)
|
||||
|
||||
|
||||
data = dg.getMessage()
|
||||
compressed = compressString(data, 9)
|
||||
key = self.generate_key(100)
|
||||
fixed = ''.join(chr((i ^ (5 * i + 7)) % ((i + 6) * 10)) for i in xrange(28))
|
||||
rc4.rc4_setkey(key + fixed)
|
||||
data = rc4.rc4(compressed)
|
||||
data = rc4.rc4(compressed)
|
||||
return key + data
|
||||
|
||||
|
||||
# 1. Make the NRI
|
||||
if args.make_nri:
|
||||
pkg = StridePackager('built/stride.dist')
|
||||
|
||||
|
||||
pkg.add_source_dir('otp')
|
||||
pkg.add_source_dir('toontown')
|
||||
|
||||
pkg.add_data_file('NiraiStart')
|
||||
|
||||
|
||||
pkg.generate_niraidata()
|
||||
pkg.write_out()
|
||||
|
||||
|
||||
# 2. Compile CXX stuff
|
||||
if args.compile_cxx:
|
||||
sys.path.append('../../../N2')
|
||||
from niraitools import NiraiCompiler
|
||||
|
||||
|
||||
compiler = NiraiCompiler('stride.exe', r'"C:\\Users\\Usuario\\workspace\\nirai-panda3d\\thirdparty\\win-libs-vc10"',
|
||||
libs=set(glob.glob('libpandadna/libpandadna.dir/Release/*.obj')))
|
||||
compiler.add_nirai_files()
|
||||
compiler.add_source('src/stride.cxx')
|
||||
|
||||
|
||||
compiler.run()
|
||||
|
|
|
@ -18,55 +18,55 @@ int niraicall_onPreStart(int argc, char* argv[])
|
|||
{
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
int niraicall_onLoadGameData()
|
||||
{
|
||||
fstream gd;
|
||||
|
||||
|
||||
// Open the file
|
||||
gd.open("stride.dist", ios_base::in | ios_base::binary);
|
||||
if (!gd.is_open())
|
||||
{
|
||||
std::cerr << "unable to open game file" << std::endl;
|
||||
std::cerr << "Unable to open game file!" << std::endl;
|
||||
return 1;
|
||||
}
|
||||
|
||||
// Check the header
|
||||
char* read_header = new char[header_size];
|
||||
gd.read(read_header, header_size);
|
||||
|
||||
|
||||
if (memcmp(header, read_header, header_size))
|
||||
{
|
||||
std::cerr << "invalid header" << std::endl;
|
||||
std::cerr << "Invalid header" << std::endl;
|
||||
return 1;
|
||||
}
|
||||
|
||||
|
||||
delete[] read_header;
|
||||
|
||||
// Extract the key
|
||||
char* key = new char[keysize + fixedsize];
|
||||
char* fixed = new char[keysize];
|
||||
|
||||
|
||||
for (int i = 0; i < fixedsize; ++i)
|
||||
fixed[i] = (i ^ (5 * i + 7)) % ((i + 6) * 10);
|
||||
|
||||
|
||||
gd.read(key, keysize);
|
||||
memcpy(&key[keysize], fixed, fixedsize);
|
||||
|
||||
std::stringstream ss;
|
||||
ss << gd.rdbuf();
|
||||
gd.close();
|
||||
|
||||
|
||||
// Decrypt
|
||||
std::string rawdata = ss.str();
|
||||
std::string decrypted_data = rc4(rawdata.c_str(), key, rawdata.size(),
|
||||
keysize + fixedsize);
|
||||
delete[] key;
|
||||
delete[] fixed;
|
||||
|
||||
|
||||
// Decompress and read
|
||||
std::string decompressed = decompress_string(decrypted_data);
|
||||
|
||||
|
||||
Datagram dg(decompressed);
|
||||
DatagramIterator dgi(dg);
|
||||
|
||||
|
@ -80,34 +80,34 @@ int niraicall_onLoadGameData()
|
|||
module = dgi.get_string();
|
||||
size = dgi.get_int32();
|
||||
data = dgi.extract_bytes(abs(size));
|
||||
|
||||
|
||||
char* name = new char[module.size() + 1];
|
||||
memcpy(name, module.c_str(), module.size());
|
||||
memset(&name[module.size()], 0, 1);
|
||||
|
||||
|
||||
unsigned char* code = new unsigned char[data.size()];
|
||||
memcpy(code, data.c_str(), data.size());
|
||||
|
||||
|
||||
_frozen fz;
|
||||
fz.name = name;
|
||||
fz.code = code;
|
||||
fz.size = size;
|
||||
|
||||
|
||||
memcpy(&fzns[i], &fz, sizeof(_frozen));
|
||||
}
|
||||
|
||||
|
||||
nassertd(dgi.get_remaining_size() == 0)
|
||||
{
|
||||
std::cerr << "corrupted data" << std::endl;
|
||||
std::cerr << "Corrupted data!" << std::endl;
|
||||
return 1;
|
||||
}
|
||||
|
||||
|
||||
memset(&fzns[num_modules], 0, sizeof(_frozen));
|
||||
PyImport_FrozenModules = fzns;
|
||||
|
||||
// libpandadna
|
||||
init_libpandadna();
|
||||
initlibpandadna();
|
||||
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
|
|
@ -211,14 +211,14 @@ class OTPClientRepository(ClientRepositoryBase):
|
|||
|
||||
def hasPlayToken():
|
||||
return self.playToken != None
|
||||
|
||||
|
||||
def readDCFile(self, dcFileNames=None):
|
||||
dcFile = self.getDcFile()
|
||||
dcFile.clear()
|
||||
self.dclassesByName = {}
|
||||
self.dclassesByNumber = {}
|
||||
self.hashVal = 0
|
||||
|
||||
|
||||
if isinstance(dcFileNames, types.StringTypes):
|
||||
# If we were given a single string, make it a list.
|
||||
dcFileNames = [dcFileNames]
|
||||
|
@ -229,13 +229,13 @@ class OTPClientRepository(ClientRepositoryBase):
|
|||
# For Nirai
|
||||
readResult = dcFile.read(dcStream, '__dc__')
|
||||
del __builtin__.dcStream
|
||||
|
||||
|
||||
except NameError:
|
||||
readResult = dcFile.readAll()
|
||||
|
||||
|
||||
if not readResult:
|
||||
self.notify.error("Could not read dc file.")
|
||||
|
||||
|
||||
else:
|
||||
searchPath = getModelPath().getValue()
|
||||
for dcFileName in dcFileNames:
|
||||
|
|
|
@ -34,7 +34,7 @@ args = parser.parse_args()
|
|||
|
||||
for prc in args.config:
|
||||
loadPrcFile(prc)
|
||||
|
||||
|
||||
if os.path.isfile('dependencies/config/local.prc'):
|
||||
loadPrcFile('dependencies/config/local.prc')
|
||||
|
||||
|
|
|
@ -31,7 +31,7 @@ if __debug__:
|
|||
|
||||
loadPrcFile('dependencies/config/general.prc')
|
||||
loadPrcFile('dependencies/config/release/dev.prc')
|
||||
|
||||
|
||||
if os.path.isfile('dependencies/config/local.prc'):
|
||||
loadPrcFile('dependencies/config/local.prc')
|
||||
|
||||
|
@ -106,28 +106,6 @@ loadPrcFileData('Settings: musicVol', 'audio-master-music-volume %s' % settings[
|
|||
loadPrcFileData('Settings: sfxVol', 'audio-master-sfx-volume %s' % settings['sfxVol'])
|
||||
loadPrcFileData('Settings: loadDisplay', 'load-display %s' % settings['loadDisplay'])
|
||||
|
||||
import os
|
||||
|
||||
from toontown.toonbase.ContentPacksManager import ContentPackError
|
||||
from toontown.toonbase.ContentPacksManager import ContentPacksManager
|
||||
|
||||
|
||||
contentPacksFilepath = ConfigVariableString(
|
||||
'content-packs-filepath', 'user/contentpacks/').getValue()
|
||||
contentPacksSortFilename = ConfigVariableString(
|
||||
'content-packs-sort-filename', 'sort.yaml').getValue()
|
||||
if not os.path.exists(contentPacksFilepath):
|
||||
os.makedirs(contentPacksFilepath)
|
||||
__builtin__.ContentPackError = ContentPackError
|
||||
__builtin__.contentPacksMgr = ContentPacksManager(
|
||||
filepath=contentPacksFilepath, sortFilename=contentPacksSortFilename)
|
||||
contentPacksMgr.applyAll()
|
||||
|
||||
languagePack = settings['language'].lower() + '.mf'
|
||||
|
||||
if contentPacksMgr.isApplicable(languagePack):
|
||||
contentPacksMgr.applyMultifile(languagePack)
|
||||
|
||||
import time
|
||||
import sys
|
||||
import random
|
||||
|
|
|
@ -29,7 +29,7 @@ def rejectConfig(issue, securityIssue=True, retarded=True):
|
|||
print '"Either down\'s or autism"\n - JohnnyDaPirate, 2015'
|
||||
print 'Go fix that!'
|
||||
exit()
|
||||
|
||||
|
||||
def entropy(string):
|
||||
prob = [float(string.count(c)) / len(string) for c in dict.fromkeys(list(string))]
|
||||
entropy = -sum([p * math.log(p) / math.log(2.0) for p in prob])
|
||||
|
@ -45,10 +45,10 @@ accountServerHashAlgo = config.GetString('account-server-hash-algo', 'sha512')
|
|||
if accountDBType == 'remote':
|
||||
if accountServerSecret == 'dev':
|
||||
rejectConfig('you have not changed the secret in config/local.prc')
|
||||
|
||||
|
||||
if len(accountServerSecret) < 16:
|
||||
rejectConfig('the secret is too small! Make it 16+ bytes', retarded=False)
|
||||
|
||||
|
||||
secretLength = len(accountServerSecret)
|
||||
ideal = entropyIdeal(secretLength) / 2
|
||||
entropy = entropy(accountServerSecret)
|
||||
|
@ -56,11 +56,11 @@ if accountDBType == 'remote':
|
|||
rejectConfig('the secret entropy is too low! For %d bytes,'
|
||||
' it should be %d. Currently it is %d' % (secretLength, ideal, entropy),
|
||||
retarded=False)
|
||||
|
||||
|
||||
hashAlgo = getattr(hashlib, accountServerHashAlgo, None)
|
||||
if not hashAlgo:
|
||||
rejectConfig('%s is not a valid hash algo' % accountServerHashAlgo, securityIssue=False)
|
||||
|
||||
|
||||
hashSize = len(hashAlgo('').digest())
|
||||
|
||||
minAccessLevel = config.GetInt('min-access-level', 100)
|
||||
|
@ -147,16 +147,16 @@ class AccountDB:
|
|||
|
||||
def lookup(self, data, callback):
|
||||
userId = data['userId']
|
||||
|
||||
|
||||
data['success'] = True
|
||||
data['accessLevel'] = max(data['accessLevel'], minAccessLevel)
|
||||
|
||||
|
||||
if str(userId) not in self.dbm:
|
||||
data['accountId'] = 0
|
||||
|
||||
|
||||
else:
|
||||
data['accountId'] = int(self.dbm[str(userId)])
|
||||
|
||||
|
||||
callback(data)
|
||||
return data
|
||||
|
||||
|
@ -171,7 +171,7 @@ class AccountDB:
|
|||
|
||||
class DeveloperAccountDB(AccountDB):
|
||||
notify = directNotify.newCategory('DeveloperAccountDB')
|
||||
|
||||
|
||||
def lookup(self, userId, callback):
|
||||
return AccountDB.lookup(self, {'userId': userId,
|
||||
'accessLevel': 700,
|
||||
|
@ -200,33 +200,33 @@ class RemoteAccountDB(AccountDB):
|
|||
Token format:
|
||||
The token is obfuscated a bit, but nothing too hard to read.
|
||||
Most of the security is based on the hash.
|
||||
|
||||
|
||||
I. Data contained in a token:
|
||||
A json-encoded dict, which contains timestamp, userid and extra info
|
||||
|
||||
|
||||
II. Token format
|
||||
X = BASE64(ROT13(DATA)[::-1])
|
||||
H = HASH(X)[::-1]
|
||||
Token = BASE64(H + X)
|
||||
'''
|
||||
|
||||
|
||||
try:
|
||||
token = token.decode('base64')
|
||||
hash, token = token[:hashSize], token[hashSize:]
|
||||
|
||||
|
||||
correctHash = hashAlgo(token + accountServerSecret).digest()
|
||||
if len(hash) != len(correctHash):
|
||||
raise ValueError('invalid hash')
|
||||
|
||||
raise ValueError('Invalid hash.')
|
||||
|
||||
value = 0
|
||||
for x, y in zip(hash[::-1], correctHash):
|
||||
value |= ord(x) ^ ord(y)
|
||||
|
||||
|
||||
if value:
|
||||
raise ValueError('invalid hash')
|
||||
|
||||
raise ValueError('Invalid hash.')
|
||||
|
||||
token = json.loads(token.decode('base64')[::-1].decode('rot13'))
|
||||
|
||||
|
||||
except:
|
||||
resp = {'success': False}
|
||||
callback(resp)
|
||||
|
@ -293,12 +293,12 @@ class LoginAccountFSM(OperationFSM):
|
|||
return
|
||||
|
||||
self.account = fields
|
||||
|
||||
|
||||
if self.notAfter:
|
||||
if self.account.get('LAST_LOGIN_TS', 0) > self.notAfter:
|
||||
self.notify.debug('Rejecting old token: %d, notAfter=%d' % (self.account.get('LAST_LOGIN_TS', 0), self.notAfter))
|
||||
return self.__handleLookup({'success': False})
|
||||
|
||||
|
||||
self.demand('SetAccount')
|
||||
|
||||
def enterCreateAccount(self):
|
||||
|
@ -1001,7 +1001,7 @@ class ClientServicesManagerUD(DistributedObjectGlobalUD):
|
|||
self.accountDB = RemoteAccountDB(self)
|
||||
else:
|
||||
self.notify.error('Invalid accountdb-type: ' + accountDBType)
|
||||
|
||||
|
||||
def killConnection(self, connId, reason):
|
||||
datagram = PyDatagram()
|
||||
datagram.addServerHeader(
|
||||
|
|
|
@ -35,7 +35,7 @@ for prc in args.config:
|
|||
|
||||
if os.path.isfile('dependencies/config/local.prc'):
|
||||
loadPrcFile('dependencies/config/local.prc')
|
||||
|
||||
|
||||
localconfig = ''
|
||||
if args.base_channel:
|
||||
localconfig += 'air-base-channel %s\n' % args.base_channel
|
||||
|
|
Loading…
Reference in a new issue