mirror of
https://github.com/Sneed-Group/Poodletooth-iLand
synced 2024-12-24 04:02:40 -06:00
Merge branch 'master' of https://github.com/stevetts/tts_src
This commit is contained in:
commit
2bbe0aaa7c
17 changed files with 113 additions and 319 deletions
|
@ -14,9 +14,9 @@ for line in prc.split('\n'):
|
|||
line = line.strip()
|
||||
if line:
|
||||
loadPrcFileData('nirai config', line)
|
||||
|
||||
|
||||
del prc
|
||||
|
||||
|
||||
# DC
|
||||
__builtin__.dcStream = StringStream()
|
||||
|
||||
|
@ -39,18 +39,18 @@ abort = False
|
|||
for mf in mfs:
|
||||
filename = 'resources/default/phase_%s.mf' % mf
|
||||
if not os.path.isfile(filename):
|
||||
print 'Phase %s not found' % filename
|
||||
print 'Phase %s not found' % filename
|
||||
abort = True
|
||||
break
|
||||
|
||||
|
||||
mf = Multifile()
|
||||
mf.openRead(filename)
|
||||
|
||||
|
||||
if not vfs.mount(mf, '../resources', 0):
|
||||
print 'Unable to mount %s' % filename
|
||||
abort = True
|
||||
break
|
||||
|
||||
|
||||
# Packs
|
||||
pack = os.environ.get('TT_STRIDE_CONTENT_PACK')
|
||||
if pack and pack != 'default':
|
||||
|
@ -63,14 +63,14 @@ if pack and pack != 'default':
|
|||
ext = os.path.splitext(name)[1]
|
||||
if ext not in ('.jpg', '.jpeg', '.ogg', '.rgb'):
|
||||
mf.removeSubfile(name)
|
||||
|
||||
|
||||
mf.flush()
|
||||
|
||||
|
||||
if not vfs.mount(mf, '../resources', 0):
|
||||
print 'Unable to mount %s' % filename
|
||||
abort = True
|
||||
break
|
||||
|
||||
|
||||
if not abort:
|
||||
# Run
|
||||
import toontown.toonbase.ClientStart
|
||||
|
|
|
@ -11,75 +11,75 @@ parser.add_argument('--make-nri', '-n', action='store_true',
|
|||
help='Generate stride NRI.')
|
||||
args = parser.parse_args()
|
||||
|
||||
# BEGIN (STRIPPED AND MODIFIED) COPY FROM niraitools.py
|
||||
# BEGIN (STRIPPED AND MODIFIED) COPY FROM niraitools.py
|
||||
class NiraiPackager:
|
||||
HEADER = 'NRI\n'
|
||||
|
||||
|
||||
def __init__(self, outfile):
|
||||
self.modules = {}
|
||||
self.outfile = outfile
|
||||
|
||||
|
||||
def __read_file(self, filename, mangler=None):
|
||||
with open(filename, 'rb') as f:
|
||||
data = f.read()
|
||||
|
||||
|
||||
base = filename.rsplit('.', 1)[0].replace('\\', '/').replace('/', '.')
|
||||
pkg = base.endswith('.__init__')
|
||||
moduleName = base.rsplit('.', 1)[0] if pkg else base
|
||||
|
||||
|
||||
name = moduleName
|
||||
if mangler is not None:
|
||||
name = mangler(name)
|
||||
|
||||
|
||||
if not name:
|
||||
return '', ('', 0)
|
||||
|
||||
|
||||
try:
|
||||
data = self.compile_module(name, data)
|
||||
|
||||
|
||||
except:
|
||||
print 'WARNING: Failed to compile', filename
|
||||
return '', ('', 0)
|
||||
|
||||
|
||||
size = len(data) * (-1 if pkg else 1)
|
||||
return name, (data, size)
|
||||
|
||||
|
||||
def compile_module(self, name, data):
|
||||
return marshal.dumps(compile(data, name, 'exec'))
|
||||
|
||||
|
||||
def add_module(self, moduleName, data, size=None, compile=False):
|
||||
if compile:
|
||||
data = self.compile_module(moduleName, data)
|
||||
|
||||
|
||||
if size is None:
|
||||
size = len(data)
|
||||
|
||||
|
||||
self.modules[moduleName] = (data, size)
|
||||
|
||||
|
||||
def add_file(self, filename, mangler=None):
|
||||
print 'Adding file', filename
|
||||
moduleName, (data, size) = self.__read_file(filename, mangler)
|
||||
if moduleName:
|
||||
moduleName = os.path.basename(filename).rsplit('.', 1)[0]
|
||||
self.add_module(moduleName, data, size)
|
||||
|
||||
|
||||
def add_directory(self, dir, mangler=None):
|
||||
print 'Adding directory', dir
|
||||
|
||||
|
||||
def _recurse_dir(dir):
|
||||
for f in os.listdir(dir):
|
||||
f = os.path.join(dir, f)
|
||||
|
||||
|
||||
if os.path.isdir(f):
|
||||
_recurse_dir(f)
|
||||
|
||||
|
||||
elif f.endswith('py'):
|
||||
moduleName, (data, size) = self.__read_file(f, mangler)
|
||||
if moduleName:
|
||||
self.add_module(moduleName, data, size)
|
||||
|
||||
|
||||
_recurse_dir(dir)
|
||||
|
||||
|
||||
def get_mangle_base(self, *path):
|
||||
return len(os.path.join(*path).rsplit('.', 1)[0].replace('\\', '/').replace('/', '.')) + 1
|
||||
|
||||
|
@ -88,112 +88,112 @@ class NiraiPackager:
|
|||
f.write(self.HEADER)
|
||||
f.write(self.process_modules())
|
||||
f.close()
|
||||
|
||||
|
||||
def generate_key(self, size=256):
|
||||
return os.urandom(size)
|
||||
|
||||
|
||||
def dump_key(self, key):
|
||||
for k in key:
|
||||
print ord(k),
|
||||
|
||||
|
||||
print
|
||||
|
||||
|
||||
def process_modules(self):
|
||||
# Pure virtual
|
||||
raise NotImplementedError('process_datagram')
|
||||
|
||||
|
||||
def get_file_contents(self, filename, keysize=0):
|
||||
with open(filename, 'rb') as f:
|
||||
data = f.read()
|
||||
|
||||
|
||||
if keysize:
|
||||
key = self.generate_key(keysize)
|
||||
rc4.rc4_setkey(key)
|
||||
data = key + rc4.rc4(data)
|
||||
|
||||
|
||||
return data
|
||||
# END COPY FROM niraitools.py
|
||||
# END COPY FROM niraitools.py
|
||||
|
||||
class StridePackager(NiraiPackager):
|
||||
HEADER = 'STRIDETT'
|
||||
BASEDIR = '..' + os.sep
|
||||
|
||||
|
||||
def __init__(self, outfile):
|
||||
NiraiPackager.__init__(self, outfile)
|
||||
self.__manglebase = self.get_mangle_base(self.BASEDIR)
|
||||
|
||||
|
||||
def add_source_dir(self, dir):
|
||||
self.add_directory(self.BASEDIR + dir, mangler=self.__mangler)
|
||||
|
||||
|
||||
def add_data_file(self, file):
|
||||
mb = self.get_mangle_base('data/')
|
||||
self.add_file('data/%s.py' % file, mangler=lambda x: x[mb:])
|
||||
|
||||
def __mangler(self, name):
|
||||
|
||||
def __mangler(self, name):
|
||||
if name.endswith('AI') or name.endswith('UD') or name in ('ToontownAIRepository', 'ToontownUberRepository',
|
||||
'ToontownInternalRepository'):
|
||||
if not 'NonRepeatableRandomSource' in name:
|
||||
return ''
|
||||
|
||||
return name[self.__manglebase:].strip('.')
|
||||
|
||||
|
||||
def generate_niraidata(self):
|
||||
print 'Generating niraidata'
|
||||
|
||||
|
||||
config = self.get_file_contents('../dependencies/config/release/en.prc')
|
||||
config += '\n\n' + self.get_file_contents('../dependencies/config/general.prc')
|
||||
key = self.generate_key(128)
|
||||
rc4.rc4_setkey(key)
|
||||
config = key + rc4.rc4(config)
|
||||
|
||||
|
||||
niraidata = 'CONFIG = %r' % config
|
||||
niraidata += '\nDC = %r' % self.get_file_contents('../dependencies/astron/dclass/stride.dc', 128)
|
||||
self.add_module('niraidata', niraidata, compile=True)
|
||||
|
||||
|
||||
def process_modules(self):
|
||||
with open('base.dg', 'rb') as f:
|
||||
basesize, = struct.unpack('<I', f.read(4))
|
||||
data = f.read()
|
||||
|
||||
|
||||
dg = Datagram()
|
||||
dg.addUint32(len(self.modules) + basesize)
|
||||
dg.appendData(data)
|
||||
|
||||
|
||||
for moduleName in self.modules:
|
||||
data, size = self.modules[moduleName]
|
||||
|
||||
|
||||
dg.addString(moduleName)
|
||||
dg.addInt32(size)
|
||||
dg.appendData(data)
|
||||
|
||||
|
||||
data = dg.getMessage()
|
||||
compressed = compressString(data, 9)
|
||||
key = self.generate_key(100)
|
||||
fixed = ''.join(chr((i ^ (5 * i + 7)) % ((i + 6) * 10)) for i in xrange(28))
|
||||
rc4.rc4_setkey(key + fixed)
|
||||
data = rc4.rc4(compressed)
|
||||
data = rc4.rc4(compressed)
|
||||
return key + data
|
||||
|
||||
|
||||
# 1. Make the NRI
|
||||
if args.make_nri:
|
||||
pkg = StridePackager('built/stride.dist')
|
||||
|
||||
|
||||
pkg.add_source_dir('otp')
|
||||
pkg.add_source_dir('toontown')
|
||||
|
||||
pkg.add_data_file('NiraiStart')
|
||||
|
||||
|
||||
pkg.generate_niraidata()
|
||||
pkg.write_out()
|
||||
|
||||
|
||||
# 2. Compile CXX stuff
|
||||
if args.compile_cxx:
|
||||
sys.path.append('../../../N2')
|
||||
from niraitools import NiraiCompiler
|
||||
|
||||
|
||||
compiler = NiraiCompiler('stride.exe', r'"C:\\Users\\Usuario\\workspace\\nirai-panda3d\\thirdparty\\win-libs-vc10"',
|
||||
libs=set(glob.glob('libpandadna/libpandadna.dir/Release/*.obj')))
|
||||
compiler.add_nirai_files()
|
||||
compiler.add_source('src/stride.cxx')
|
||||
|
||||
|
||||
compiler.run()
|
||||
|
|
|
@ -18,55 +18,55 @@ int niraicall_onPreStart(int argc, char* argv[])
|
|||
{
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
int niraicall_onLoadGameData()
|
||||
{
|
||||
fstream gd;
|
||||
|
||||
|
||||
// Open the file
|
||||
gd.open("stride.dist", ios_base::in | ios_base::binary);
|
||||
if (!gd.is_open())
|
||||
{
|
||||
std::cerr << "unable to open game file" << std::endl;
|
||||
std::cerr << "Unable to open game file!" << std::endl;
|
||||
return 1;
|
||||
}
|
||||
|
||||
// Check the header
|
||||
char* read_header = new char[header_size];
|
||||
gd.read(read_header, header_size);
|
||||
|
||||
|
||||
if (memcmp(header, read_header, header_size))
|
||||
{
|
||||
std::cerr << "invalid header" << std::endl;
|
||||
std::cerr << "Invalid header" << std::endl;
|
||||
return 1;
|
||||
}
|
||||
|
||||
|
||||
delete[] read_header;
|
||||
|
||||
// Extract the key
|
||||
char* key = new char[keysize + fixedsize];
|
||||
char* fixed = new char[keysize];
|
||||
|
||||
|
||||
for (int i = 0; i < fixedsize; ++i)
|
||||
fixed[i] = (i ^ (5 * i + 7)) % ((i + 6) * 10);
|
||||
|
||||
|
||||
gd.read(key, keysize);
|
||||
memcpy(&key[keysize], fixed, fixedsize);
|
||||
|
||||
std::stringstream ss;
|
||||
ss << gd.rdbuf();
|
||||
gd.close();
|
||||
|
||||
|
||||
// Decrypt
|
||||
std::string rawdata = ss.str();
|
||||
std::string decrypted_data = rc4(rawdata.c_str(), key, rawdata.size(),
|
||||
keysize + fixedsize);
|
||||
delete[] key;
|
||||
delete[] fixed;
|
||||
|
||||
|
||||
// Decompress and read
|
||||
std::string decompressed = decompress_string(decrypted_data);
|
||||
|
||||
|
||||
Datagram dg(decompressed);
|
||||
DatagramIterator dgi(dg);
|
||||
|
||||
|
@ -80,34 +80,34 @@ int niraicall_onLoadGameData()
|
|||
module = dgi.get_string();
|
||||
size = dgi.get_int32();
|
||||
data = dgi.extract_bytes(abs(size));
|
||||
|
||||
|
||||
char* name = new char[module.size() + 1];
|
||||
memcpy(name, module.c_str(), module.size());
|
||||
memset(&name[module.size()], 0, 1);
|
||||
|
||||
|
||||
unsigned char* code = new unsigned char[data.size()];
|
||||
memcpy(code, data.c_str(), data.size());
|
||||
|
||||
|
||||
_frozen fz;
|
||||
fz.name = name;
|
||||
fz.code = code;
|
||||
fz.size = size;
|
||||
|
||||
|
||||
memcpy(&fzns[i], &fz, sizeof(_frozen));
|
||||
}
|
||||
|
||||
|
||||
nassertd(dgi.get_remaining_size() == 0)
|
||||
{
|
||||
std::cerr << "corrupted data" << std::endl;
|
||||
std::cerr << "Corrupted data!" << std::endl;
|
||||
return 1;
|
||||
}
|
||||
|
||||
|
||||
memset(&fzns[num_modules], 0, sizeof(_frozen));
|
||||
PyImport_FrozenModules = fzns;
|
||||
|
||||
// libpandadna
|
||||
init_libpandadna();
|
||||
initlibpandadna();
|
||||
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
|
4
dependencies/config/general.prc
vendored
4
dependencies/config/general.prc
vendored
|
@ -23,10 +23,6 @@ texture-anisotropic-degree 16
|
|||
# Preferences:
|
||||
preferences-filename user/preferences.json
|
||||
|
||||
# Content packs:
|
||||
content-packs-filepath user/contentpacks/
|
||||
content-packs-sort-filename sort.yaml
|
||||
|
||||
# Backups:
|
||||
backups-filepath dependencies/backups/
|
||||
backups-extension .json
|
||||
|
|
|
@ -23,7 +23,7 @@ if %INPUT%==1 (
|
|||
set TTS_GAMESERVER=127.0.0.1
|
||||
) else if %INPUT%==3 (
|
||||
echo.
|
||||
set /P TTS_GAMESERVER=Gameserver:
|
||||
set /P TTS_GAMESERVER=Gameserver:
|
||||
) else (
|
||||
goto selection
|
||||
)
|
||||
|
@ -48,6 +48,8 @@ echo ppython: "dependencies/panda/python/ppython.exe"
|
|||
|
||||
if %INPUT%==2 (
|
||||
echo Username: %ttsUsername%
|
||||
) else if %INPUT%==4 (
|
||||
echo Username: %ttsUsername%
|
||||
) else (
|
||||
echo Username: %TTS_PLAYCOOKIE%
|
||||
)
|
||||
|
|
|
@ -211,14 +211,14 @@ class OTPClientRepository(ClientRepositoryBase):
|
|||
|
||||
def hasPlayToken():
|
||||
return self.playToken != None
|
||||
|
||||
|
||||
def readDCFile(self, dcFileNames=None):
|
||||
dcFile = self.getDcFile()
|
||||
dcFile.clear()
|
||||
self.dclassesByName = {}
|
||||
self.dclassesByNumber = {}
|
||||
self.hashVal = 0
|
||||
|
||||
|
||||
if isinstance(dcFileNames, types.StringTypes):
|
||||
# If we were given a single string, make it a list.
|
||||
dcFileNames = [dcFileNames]
|
||||
|
@ -229,13 +229,13 @@ class OTPClientRepository(ClientRepositoryBase):
|
|||
# For Nirai
|
||||
readResult = dcFile.read(dcStream, '__dc__')
|
||||
del __builtin__.dcStream
|
||||
|
||||
|
||||
except NameError:
|
||||
readResult = dcFile.readAll()
|
||||
|
||||
|
||||
if not readResult:
|
||||
self.notify.error("Could not read dc file.")
|
||||
|
||||
|
||||
else:
|
||||
searchPath = getModelPath().getValue()
|
||||
for dcFileName in dcFileNames:
|
||||
|
|
|
@ -34,7 +34,7 @@ args = parser.parse_args()
|
|||
|
||||
for prc in args.config:
|
||||
loadPrcFile(prc)
|
||||
|
||||
|
||||
if os.path.isfile('dependencies/config/local.prc'):
|
||||
loadPrcFile('dependencies/config/local.prc')
|
||||
|
||||
|
|
|
@ -1,11 +1,9 @@
|
|||
from pandac.PandaModules import Vec4
|
||||
|
||||
from toontown.safezone.DDSafeZoneLoader import DDSafeZoneLoader
|
||||
from toontown.town.DDTownLoader import DDTownLoader
|
||||
from toontown.toonbase import ToontownGlobals
|
||||
from toontown.hood.ToonHood import ToonHood
|
||||
|
||||
|
||||
class DDHood(ToonHood):
|
||||
notify = directNotify.newCategory('DDHood')
|
||||
|
||||
|
@ -16,7 +14,7 @@ class DDHood(ToonHood):
|
|||
SKY_FILE = 'phase_3.5/models/props/BR_sky'
|
||||
SPOOKY_SKY_FILE = 'phase_3.5/models/props/BR_sky'
|
||||
TITLE_COLOR = (0.8, 0.6, 0.5, 1.0)
|
||||
|
||||
underwaterColor = Vec4(0.0, 0.0, 0.6, 1.0)
|
||||
HOLIDAY_DNA = {
|
||||
ToontownGlobals.WINTER_DECORATIONS: ['phase_6/dna/winter_storage_DD.pdna'],
|
||||
ToontownGlobals.WACKY_WINTER_DECORATIONS: ['phase_6/dna/winter_storage_DD.pdna'],
|
||||
|
@ -26,18 +24,6 @@ class DDHood(ToonHood):
|
|||
def __init__(self, parentFSM, doneEvent, dnaStore, hoodId):
|
||||
ToonHood.__init__(self, parentFSM, doneEvent, dnaStore, hoodId)
|
||||
|
||||
# Load content pack ambience settings:
|
||||
ambience = contentPacksMgr.getAmbience('donalds-dock')
|
||||
|
||||
color = ambience.get('underwater-color')
|
||||
if color is not None:
|
||||
try:
|
||||
self.underwaterColor = Vec4(color['r'], color['g'], color['b'], color['a'])
|
||||
except Exception, e:
|
||||
raise ContentPackError(e)
|
||||
elif self.underwaterColor is None:
|
||||
self.underwaterColor = Vec4(0, 0, 0.6, 1)
|
||||
|
||||
def load(self):
|
||||
ToonHood.load(self)
|
||||
|
||||
|
|
|
@ -1,10 +1,8 @@
|
|||
from pandac.PandaModules import Vec4
|
||||
|
||||
from toontown.safezone.OZSafeZoneLoader import OZSafeZoneLoader
|
||||
from toontown.toonbase import ToontownGlobals
|
||||
from toontown.hood.ToonHood import ToonHood
|
||||
|
||||
|
||||
class OZHood(ToonHood):
|
||||
notify = directNotify.newCategory('OZHood')
|
||||
|
||||
|
@ -14,22 +12,11 @@ class OZHood(ToonHood):
|
|||
SKY_FILE = 'phase_3.5/models/props/TT_sky'
|
||||
SPOOKY_SKY_FILE = 'phase_3.5/models/props/BR_sky'
|
||||
TITLE_COLOR = (1.0, 0.5, 0.4, 1.0)
|
||||
underwaterColor = Vec4(0.0, 0.0, 0.6, 1.0)
|
||||
|
||||
def __init__(self, parentFSM, doneEvent, dnaStore, hoodId):
|
||||
ToonHood.__init__(self, parentFSM, doneEvent, dnaStore, hoodId)
|
||||
|
||||
# Load content pack ambience settings:
|
||||
ambience = contentPacksMgr.getAmbience('outdoor-zone')
|
||||
|
||||
color = ambience.get('underwater-color')
|
||||
if color is not None:
|
||||
try:
|
||||
self.underwaterColor = Vec4(color['r'], color['g'], color['b'], color['a'])
|
||||
except Exception, e:
|
||||
raise ContentPackError(e)
|
||||
elif self.underwaterColor is None:
|
||||
self.underwaterColor = Vec4(0, 0, 0.6, 1)
|
||||
|
||||
def load(self):
|
||||
ToonHood.load(self)
|
||||
|
||||
|
|
|
@ -11,7 +11,6 @@ from toontown.cogdominium import CogdoInterior
|
|||
from toontown.toon.Toon import teleportDebug
|
||||
from toontown.hood import SkyUtil
|
||||
|
||||
|
||||
class ToonHood(Hood):
|
||||
notify = directNotify.newCategory('ToonHood')
|
||||
|
||||
|
@ -22,7 +21,6 @@ class ToonHood(Hood):
|
|||
SKY_FILE = None
|
||||
SPOOKY_SKY_FILE = None
|
||||
TITLE_COLOR = None
|
||||
|
||||
HOLIDAY_DNA = {}
|
||||
|
||||
def __init__(self, parentFSM, doneEvent, dnaStore, hoodId):
|
||||
|
@ -54,18 +52,6 @@ class ToonHood(Hood):
|
|||
State.State('final', self.enterFinal, self.exitFinal, [])], 'start', 'final')
|
||||
self.fsm.enterInitialState()
|
||||
|
||||
# Load content pack ambience settings:
|
||||
ambience = contentPacksMgr.getAmbience('general')
|
||||
|
||||
color = ambience.get('underwater-color')
|
||||
if color is not None:
|
||||
try:
|
||||
self.underwaterColor = Vec4(color['r'], color['g'], color['b'], color['a'])
|
||||
except Exception, e:
|
||||
raise ContentPackError(e)
|
||||
else:
|
||||
self.underwaterColor = None
|
||||
|
||||
# Until we cleanup Hood, we will need to define some variables
|
||||
self.id = self.ID
|
||||
self.storageDNAFile = self.STORAGE_DNA
|
||||
|
|
|
@ -31,7 +31,7 @@ if __debug__:
|
|||
|
||||
loadPrcFile('dependencies/config/general.prc')
|
||||
loadPrcFile('dependencies/config/release/dev.prc')
|
||||
|
||||
|
||||
if os.path.isfile('dependencies/config/local.prc'):
|
||||
loadPrcFile('dependencies/config/local.prc')
|
||||
|
||||
|
@ -106,28 +106,6 @@ loadPrcFileData('Settings: musicVol', 'audio-master-music-volume %s' % settings[
|
|||
loadPrcFileData('Settings: sfxVol', 'audio-master-sfx-volume %s' % settings['sfxVol'])
|
||||
loadPrcFileData('Settings: loadDisplay', 'load-display %s' % settings['loadDisplay'])
|
||||
|
||||
import os
|
||||
|
||||
from toontown.toonbase.ContentPacksManager import ContentPackError
|
||||
from toontown.toonbase.ContentPacksManager import ContentPacksManager
|
||||
|
||||
|
||||
contentPacksFilepath = ConfigVariableString(
|
||||
'content-packs-filepath', 'user/contentpacks/').getValue()
|
||||
contentPacksSortFilename = ConfigVariableString(
|
||||
'content-packs-sort-filename', 'sort.yaml').getValue()
|
||||
if not os.path.exists(contentPacksFilepath):
|
||||
os.makedirs(contentPacksFilepath)
|
||||
__builtin__.ContentPackError = ContentPackError
|
||||
__builtin__.contentPacksMgr = ContentPacksManager(
|
||||
filepath=contentPacksFilepath, sortFilename=contentPacksSortFilename)
|
||||
contentPacksMgr.applyAll()
|
||||
|
||||
languagePack = settings['language'].lower() + '.mf'
|
||||
|
||||
if contentPacksMgr.isApplicable(languagePack):
|
||||
contentPacksMgr.applyMultifile(languagePack)
|
||||
|
||||
import time
|
||||
import sys
|
||||
import random
|
||||
|
|
|
@ -24,6 +24,6 @@ else:
|
|||
print response['reason']
|
||||
else:
|
||||
os.environ['TTS_PLAYCOOKIE'] = response['token']
|
||||
|
||||
|
||||
# Start the game:
|
||||
import toontown.toonbase.ClientStart
|
||||
|
|
|
@ -1,138 +0,0 @@
|
|||
from direct.directnotify.DirectNotifyGlobal import directNotify
|
||||
import fnmatch
|
||||
import os
|
||||
import yaml
|
||||
from panda3d.core import Multifile, Filename, VirtualFileSystem
|
||||
|
||||
APPLICABLE_FILE_PATTERNS = ('*.mf', 'ambience.yaml')
|
||||
CONTENT_EXT_WHITELIST = ('.jpg', '.jpeg', '.rgb', '.png', '.ogg', '.ttf')
|
||||
|
||||
class ContentPackError(Exception):
|
||||
pass
|
||||
|
||||
class ContentPacksManager:
|
||||
notify = directNotify.newCategory('ContentPacksManager')
|
||||
notify.setInfo(True)
|
||||
|
||||
def __init__(self, filepath='user/contentpacks/', sortFilename='sort.yaml'):
|
||||
self.filepath = filepath
|
||||
self.sortFilename = os.path.join(self.filepath, sortFilename)
|
||||
|
||||
if __debug__:
|
||||
self.mountPoint = '../resources'
|
||||
else:
|
||||
self.mountPoint = '/'
|
||||
|
||||
self.vfs = VirtualFileSystem.getGlobalPtr()
|
||||
|
||||
self.sort = []
|
||||
self.ambience = {}
|
||||
|
||||
def isApplicable(self, filename):
|
||||
"""
|
||||
Returns whether or not the specified file is applicable.
|
||||
"""
|
||||
# Does this file exist?
|
||||
if not os.path.exists(filename):
|
||||
return False
|
||||
|
||||
# Does this file match one of the applicable file patterns?
|
||||
basename = os.path.basename(filename)
|
||||
for pattern in APPLICABLE_FILE_PATTERNS:
|
||||
if fnmatch.fnmatch(basename, pattern):
|
||||
return True
|
||||
|
||||
return False
|
||||
|
||||
def applyMultifile(self, filename):
|
||||
"""
|
||||
Apply the specified multifile.
|
||||
"""
|
||||
mf = Multifile()
|
||||
mf.openReadWrite(Filename(filename))
|
||||
|
||||
# Discard content with non-whitelisted extensions:
|
||||
for subfileName in mf.getSubfileNames():
|
||||
ext = os.path.splitext(subfileName)[1]
|
||||
if ext not in CONTENT_EXT_WHITELIST:
|
||||
mf.removeSubfile(subfileName)
|
||||
|
||||
self.vfs.mount(mf, self.mountPoint, 0)
|
||||
|
||||
def applyAmbience(self, filename):
|
||||
"""
|
||||
Apply the specified ambience configuration file.
|
||||
"""
|
||||
with open(os.path.join(self.filepath, filename), 'r') as f:
|
||||
self.ambience.update(yaml.load(f) or {})
|
||||
|
||||
def apply(self, filename):
|
||||
"""
|
||||
Apply the specified content pack file.
|
||||
"""
|
||||
self.notify.info('Applying %s...' % filename)
|
||||
basename = os.path.basename(filename)
|
||||
if basename.endswith('.mf'):
|
||||
self.applyMultifile(os.path.join(self.filepath, filename))
|
||||
elif basename == 'ambience.yaml':
|
||||
self.applyAmbience(filename)
|
||||
|
||||
def applyAll(self):
|
||||
"""
|
||||
Using the sort configuration, recursively apply all applicable content
|
||||
pack files under the configured content packs directory.
|
||||
"""
|
||||
# First, read the sort configuration:
|
||||
self.readSortConfig()
|
||||
|
||||
# Next, apply the sorted files:
|
||||
for filename in self.sort[:]:
|
||||
if self.isApplicable(os.path.join(self.filepath, filename)):
|
||||
self.apply(filename)
|
||||
else:
|
||||
self.notify.warning('Invalidating %s...' % filename)
|
||||
self.sort.remove(filename)
|
||||
|
||||
# Apply the non-sorted files:
|
||||
for root, _, filenames in os.walk(self.filepath):
|
||||
root = root[len(self.filepath):]
|
||||
for filename in filenames:
|
||||
filename = os.path.join(root, filename).replace('\\', '/')
|
||||
|
||||
# Ensure this file isn't sorted:
|
||||
if filename in self.sort:
|
||||
continue
|
||||
|
||||
# Ensure this file is applicable:
|
||||
if not self.isApplicable(os.path.join(self.filepath, filename)):
|
||||
continue
|
||||
|
||||
# Apply this file, and add it to the sort configuration:
|
||||
self.apply(filename)
|
||||
self.sort.append(filename)
|
||||
|
||||
# Finally, write the new sort configuration:
|
||||
self.writeSortConfig()
|
||||
|
||||
def readSortConfig(self):
|
||||
"""
|
||||
Read the sort configuration.
|
||||
"""
|
||||
if not os.path.exists(self.sortFilename):
|
||||
return
|
||||
with open(self.sortFilename, 'r') as f:
|
||||
self.sort = yaml.load(f) or []
|
||||
|
||||
def writeSortConfig(self):
|
||||
"""
|
||||
Write the sort configuration to disk.
|
||||
"""
|
||||
with open(self.sortFilename, 'w') as f:
|
||||
for filename in self.sort:
|
||||
f.write('- %s\n' % filename)
|
||||
|
||||
def getAmbience(self, group):
|
||||
"""
|
||||
Returns the ambience configurations for the specified group.
|
||||
"""
|
||||
return self.ambience.get(group, {})
|
|
@ -29,7 +29,7 @@ def rejectConfig(issue, securityIssue=True, retarded=True):
|
|||
print '"Either down\'s or autism"\n - JohnnyDaPirate, 2015'
|
||||
print 'Go fix that!'
|
||||
exit()
|
||||
|
||||
|
||||
def entropy(string):
|
||||
prob = [float(string.count(c)) / len(string) for c in dict.fromkeys(list(string))]
|
||||
entropy = -sum([p * math.log(p) / math.log(2.0) for p in prob])
|
||||
|
@ -45,10 +45,10 @@ accountServerHashAlgo = config.GetString('account-server-hash-algo', 'sha512')
|
|||
if accountDBType == 'remote':
|
||||
if accountServerSecret == 'dev':
|
||||
rejectConfig('you have not changed the secret in config/local.prc')
|
||||
|
||||
|
||||
if len(accountServerSecret) < 16:
|
||||
rejectConfig('the secret is too small! Make it 16+ bytes', retarded=False)
|
||||
|
||||
|
||||
secretLength = len(accountServerSecret)
|
||||
ideal = entropyIdeal(secretLength) / 2
|
||||
entropy = entropy(accountServerSecret)
|
||||
|
@ -56,11 +56,11 @@ if accountDBType == 'remote':
|
|||
rejectConfig('the secret entropy is too low! For %d bytes,'
|
||||
' it should be %d. Currently it is %d' % (secretLength, ideal, entropy),
|
||||
retarded=False)
|
||||
|
||||
|
||||
hashAlgo = getattr(hashlib, accountServerHashAlgo, None)
|
||||
if not hashAlgo:
|
||||
rejectConfig('%s is not a valid hash algo' % accountServerHashAlgo, securityIssue=False)
|
||||
|
||||
|
||||
hashSize = len(hashAlgo('').digest())
|
||||
|
||||
minAccessLevel = config.GetInt('min-access-level', 100)
|
||||
|
@ -147,16 +147,16 @@ class AccountDB:
|
|||
|
||||
def lookup(self, data, callback):
|
||||
userId = data['userId']
|
||||
|
||||
|
||||
data['success'] = True
|
||||
data['accessLevel'] = max(data['accessLevel'], minAccessLevel)
|
||||
|
||||
|
||||
if str(userId) not in self.dbm:
|
||||
data['accountId'] = 0
|
||||
|
||||
|
||||
else:
|
||||
data['accountId'] = int(self.dbm[str(userId)])
|
||||
|
||||
|
||||
callback(data)
|
||||
return data
|
||||
|
||||
|
@ -171,7 +171,7 @@ class AccountDB:
|
|||
|
||||
class DeveloperAccountDB(AccountDB):
|
||||
notify = directNotify.newCategory('DeveloperAccountDB')
|
||||
|
||||
|
||||
def lookup(self, userId, callback):
|
||||
return AccountDB.lookup(self, {'userId': userId,
|
||||
'accessLevel': 700,
|
||||
|
@ -200,33 +200,33 @@ class RemoteAccountDB(AccountDB):
|
|||
Token format:
|
||||
The token is obfuscated a bit, but nothing too hard to read.
|
||||
Most of the security is based on the hash.
|
||||
|
||||
|
||||
I. Data contained in a token:
|
||||
A json-encoded dict, which contains timestamp, userid and extra info
|
||||
|
||||
|
||||
II. Token format
|
||||
X = BASE64(ROT13(DATA)[::-1])
|
||||
H = HASH(X)[::-1]
|
||||
Token = BASE64(H + X)
|
||||
'''
|
||||
|
||||
|
||||
try:
|
||||
token = token.decode('base64')
|
||||
hash, token = token[:hashSize], token[hashSize:]
|
||||
|
||||
|
||||
correctHash = hashAlgo(token + accountServerSecret).digest()
|
||||
if len(hash) != len(correctHash):
|
||||
raise ValueError('invalid hash')
|
||||
|
||||
raise ValueError('Invalid hash.')
|
||||
|
||||
value = 0
|
||||
for x, y in zip(hash[::-1], correctHash):
|
||||
value |= ord(x) ^ ord(y)
|
||||
|
||||
|
||||
if value:
|
||||
raise ValueError('invalid hash')
|
||||
|
||||
raise ValueError('Invalid hash.')
|
||||
|
||||
token = json.loads(token.decode('base64')[::-1].decode('rot13'))
|
||||
|
||||
|
||||
except:
|
||||
resp = {'success': False}
|
||||
callback(resp)
|
||||
|
@ -293,12 +293,12 @@ class LoginAccountFSM(OperationFSM):
|
|||
return
|
||||
|
||||
self.account = fields
|
||||
|
||||
|
||||
if self.notAfter:
|
||||
if self.account.get('LAST_LOGIN_TS', 0) > self.notAfter:
|
||||
self.notify.debug('Rejecting old token: %d, notAfter=%d' % (self.account.get('LAST_LOGIN_TS', 0), self.notAfter))
|
||||
return self.__handleLookup({'success': False})
|
||||
|
||||
|
||||
self.demand('SetAccount')
|
||||
|
||||
def enterCreateAccount(self):
|
||||
|
@ -1001,7 +1001,7 @@ class ClientServicesManagerUD(DistributedObjectGlobalUD):
|
|||
self.accountDB = RemoteAccountDB(self)
|
||||
else:
|
||||
self.notify.error('Invalid accountdb-type: ' + accountDBType)
|
||||
|
||||
|
||||
def killConnection(self, connId, reason):
|
||||
datagram = PyDatagram()
|
||||
datagram.addServerHeader(
|
||||
|
|
|
@ -35,7 +35,7 @@ for prc in args.config:
|
|||
|
||||
if os.path.isfile('dependencies/config/local.prc'):
|
||||
loadPrcFile('dependencies/config/local.prc')
|
||||
|
||||
|
||||
localconfig = ''
|
||||
if args.base_channel:
|
||||
localconfig += 'air-base-channel %s\n' % args.base_channel
|
||||
|
|
3
user/contentpacks/.gitignore
vendored
3
user/contentpacks/.gitignore
vendored
|
@ -1,3 +0,0 @@
|
|||
*
|
||||
!.gitignore
|
||||
!sort.yaml
|
Loading…
Reference in a new issue