540 lines
21 KiB
Python
540 lines
21 KiB
Python
##############################################################################
|
|
#
|
|
# Copyright (c) 2004-2008 Tiny SPRL (http://tiny.be) All Rights Reserved.
|
|
#
|
|
# $Id$
|
|
#
|
|
# WARNING: This program as such is intended to be used by professional
|
|
# programmers who take the whole responsability of assessing all potential
|
|
# consequences resulting from its eventual inadequacies and bugs
|
|
# End users who are looking for a ready-to-use solution with commercial
|
|
# garantees and support are strongly adviced to contract a Free Software
|
|
# Service Company
|
|
#
|
|
# This program is Free Software; you can redistribute it and/or
|
|
# modify it under the terms of the GNU General Public License
|
|
# as published by the Free Software Foundation; either version 2
|
|
# of the License, or (at your option) any later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this program; if not, write to the Free Software
|
|
# Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
|
|
###############################################################################
|
|
|
|
import os
|
|
from os.path import join
|
|
import fnmatch
|
|
import csv, xml.dom, re
|
|
import osv, tools, pooler
|
|
import ir
|
|
import netsvc
|
|
from tools.misc import UpdateableStr
|
|
import inspect
|
|
import mx.DateTime as mxdt
|
|
|
|
class UNIX_LINE_TERMINATOR(csv.excel):
|
|
lineterminator = '\n'
|
|
|
|
csv.register_dialect("UNIX", UNIX_LINE_TERMINATOR)
|
|
|
|
#
|
|
# TODO: a caching method
|
|
#
|
|
def translate(cr, name, source_type, lang, source=None):
|
|
if source and name:
|
|
cr.execute('select value from ir_translation where lang=%s and type=%s and name=%s and src=%s', (lang, source_type, str(name), source))
|
|
elif name:
|
|
cr.execute('select value from ir_translation where lang=%s and type=%s and name=%s', (lang, source_type, str(name)))
|
|
elif source:
|
|
cr.execute('select value from ir_translation where lang=%s and type=%s and src=%s', (lang, source_type, source))
|
|
res_trans = cr.fetchone()
|
|
res = res_trans and res_trans[0] or False
|
|
return res
|
|
|
|
class GettextAlias(object):
|
|
def __call__(self, source):
|
|
frame = inspect.stack()[1][0]
|
|
cr = frame.f_locals['cr']
|
|
context = frame.f_locals.get('context', {})
|
|
lang = context.get('lang', False)
|
|
if not lang:
|
|
return source
|
|
return translate(cr, None, 'code', lang, source) or source
|
|
|
|
_ = GettextAlias()
|
|
|
|
|
|
# class to handle po files
|
|
class TinyPoFile(object):
|
|
def __init__(self, buffer):
|
|
self.buffer = buffer
|
|
|
|
def __iter__(self):
|
|
self.buffer.seek(0)
|
|
self.lines = self.buffer.readlines()
|
|
self.first = True
|
|
return self
|
|
|
|
def next(self):
|
|
def unquote(str):
|
|
return str[1:-1].replace("\\n", "\n") \
|
|
.replace('\\"', "\"")
|
|
|
|
type = name = res_id = source = trad = None
|
|
|
|
line = None
|
|
while not line:
|
|
if 0 == len(self.lines):
|
|
raise StopIteration()
|
|
line = self.lines.pop(0).strip()
|
|
|
|
while line.startswith('#'):
|
|
if line.startswith('#:'):
|
|
type, name, res_id = line[2:].strip().split(':')
|
|
line = self.lines.pop(0).strip()
|
|
if not line.startswith('msgid'):
|
|
raise Exception("malformed file")
|
|
source = line[7:-1]
|
|
line = self.lines.pop(0).strip()
|
|
if not source and self.first:
|
|
# if the source is "" and it's the first msgid, it's the special
|
|
# msgstr with the informations about the traduction and the
|
|
# traductor; we skip it
|
|
while line:
|
|
line = self.lines.pop(0).strip()
|
|
return next()
|
|
|
|
while not line.startswith('msgstr'):
|
|
if not line:
|
|
raise Exception('malformed file')
|
|
source += unquote(line)
|
|
line = self.lines.pop(0).strip()
|
|
|
|
trad = line[8:-1]
|
|
line = self.lines.pop(0).strip()
|
|
while line:
|
|
trad += unquote(line)
|
|
line = self.lines.pop(0).strip()
|
|
|
|
self.first = False
|
|
return type, name, res_id, source, trad
|
|
|
|
def write_infos(self, modules):
|
|
import release
|
|
self.buffer.write("# Translation of %(project)s.\n" \
|
|
"# This file containt the translation of the following modules:\n" \
|
|
"%(modules)s" \
|
|
"#\n" \
|
|
"msgid \"\"\n" \
|
|
"msgstr \"\"\n" \
|
|
"\"Project-Id-Version: %(project)s %(version)s\"\n" \
|
|
"\"Report-Msgid-Bugs-To: %(bugmail)s\"\n" \
|
|
"\"POT-Creation-Date: %(now)s\"\n" \
|
|
"\"PO-Revision-Date: %(now)s\"\n" \
|
|
"\"Last-Translator: <>\"\n" \
|
|
"\"Language-Team: \"\n" \
|
|
"\"MIME-Version: 1.0\"\n" \
|
|
"\"Content-Type: text/plain; charset=UTF-8\"\n" \
|
|
"\"Content-Transfer-Encoding: \"\n" \
|
|
"\"Plural-Forms: \"\n" \
|
|
"\n"
|
|
|
|
% { 'project': release.description,
|
|
'version': release.version,
|
|
'modules': reduce(lambda s, m: s + "#\t* %s\n" % m, modules, ""),
|
|
'bugmail': 'support@tinyerp.com', #TODO: use variable from release
|
|
'now': mxdt.ISO.strUTC(mxdt.ISO.DateTime.utc()),
|
|
}
|
|
)
|
|
|
|
def write(self, module, type, name, res_id, source, trad):
|
|
def quote(str):
|
|
return '"%s"' % str.replace('"','\\"') \
|
|
.replace('\n', '\\n"\n"')
|
|
|
|
self.buffer.write("#. module: %s\n" \
|
|
"#, python-format\n" \
|
|
"#: %s:%s:%s\n" \
|
|
"msgid %s\n" \
|
|
"msgstr %s\n\n" \
|
|
% (module, type, name, str(res_id), quote(source), quote(trad))
|
|
)
|
|
|
|
|
|
|
|
# Methods to export the translation file
|
|
|
|
def trans_export(lang, modules, buffer, format, dbname=None):
|
|
trans = trans_generate(lang, modules, dbname)
|
|
if format == 'csv':
|
|
writer=csv.writer(buffer, 'UNIX')
|
|
for row in trans:
|
|
writer.writerow(row)
|
|
elif format == 'po':
|
|
trans.pop(0)
|
|
writer = tools.TinyPoFile(buffer)
|
|
writer.write_infos(modules)
|
|
for module, type, name, res_id, src, trad in trans:
|
|
writer.write(module, type, name, res_id, src, trad)
|
|
else:
|
|
raise Exception(_('Bad file format'))
|
|
del trans
|
|
|
|
|
|
def trans_parse_xsl(de):
|
|
res = []
|
|
for n in [i for i in de.childNodes if (i.nodeType == i.ELEMENT_NODE)]:
|
|
if n.hasAttribute("t"):
|
|
for m in [j for j in n.childNodes if (j.nodeType == j.TEXT_NODE)]:
|
|
l = m.data.strip().replace('\n',' ')
|
|
if len(l):
|
|
res.append(l.encode("utf8"))
|
|
res.extend(trans_parse_xsl(n))
|
|
return res
|
|
|
|
def trans_parse_rml(de):
|
|
res = []
|
|
for n in [i for i in de.childNodes if (i.nodeType == i.ELEMENT_NODE)]:
|
|
for m in [j for j in n.childNodes if (j.nodeType == j.TEXT_NODE)]:
|
|
string_list = [s.replace('\n', ' ').strip() for s in re.split('\[\[.+?\]\]', m.data)]
|
|
for s in string_list:
|
|
if s:
|
|
res.append(s.encode("utf8"))
|
|
res.extend(trans_parse_rml(n))
|
|
return res
|
|
|
|
def trans_parse_view(de):
|
|
res = []
|
|
if de.hasAttribute("string"):
|
|
s = de.getAttribute('string')
|
|
if s:
|
|
res.append(s.encode("utf8"))
|
|
if de.hasAttribute("sum"):
|
|
s = de.getAttribute('sum')
|
|
if s:
|
|
res.append(s.encode("utf8"))
|
|
for n in [i for i in de.childNodes if (i.nodeType == i.ELEMENT_NODE)]:
|
|
res.extend(trans_parse_view(n))
|
|
return res
|
|
|
|
# tests whether an object is in a list of modules
|
|
def in_modules(object_name, modules):
|
|
if 'all' in modules:
|
|
return True
|
|
|
|
module_dict = {
|
|
'ir': 'base',
|
|
'res': 'base',
|
|
'workflow': 'base',
|
|
}
|
|
module = object_name.split('.')[0]
|
|
module = module_dict.get(module, module)
|
|
return module in modules
|
|
|
|
def trans_generate(lang, modules, dbname=None):
|
|
logger = netsvc.Logger()
|
|
if not dbname:
|
|
dbname=tools.config['db_name']
|
|
pool = pooler.get_pool(dbname)
|
|
trans_obj = pool.get('ir.translation')
|
|
model_data_obj = pool.get('ir.model.data')
|
|
cr = pooler.get_db(dbname).cursor()
|
|
uid = 1
|
|
l = pool.obj_pool.items()
|
|
l.sort()
|
|
|
|
query = 'SELECT name, model, res_id, module' \
|
|
' FROM ir_model_data'
|
|
if not 'all' in modules:
|
|
query += ' WHERE module IN (%s)' % ','.join(['%s']*len(modules))
|
|
query += ' ORDER BY module, model, name'
|
|
|
|
query_param = not 'all' in modules and modules or None
|
|
cr.execute(query, query_param)
|
|
|
|
#if 'all' in modules:
|
|
# cr.execute('select name,model,res_id,module from ir_model_data')
|
|
#else:
|
|
# cr.execute('select name,model,res_id,module from ir_model_data where module in ('+','.join(['%s']*len(modules))+')', modules)
|
|
|
|
|
|
_to_translate = []
|
|
def push_translation(module, type, name, id, source):
|
|
tuple = (module, type, name, id, source)
|
|
if not tuple in _to_translate:
|
|
_to_translate.append(tuple)
|
|
|
|
|
|
for (xml_name,model,res_id,module) in cr.fetchall():
|
|
xml_name = module+'.'+xml_name
|
|
obj = pool.get(model).browse(cr, uid, res_id)
|
|
if model=='ir.ui.view':
|
|
d = xml.dom.minidom.parseString(obj.arch)
|
|
for t in trans_parse_view(d.documentElement):
|
|
push_translation(module, 'view', obj.model, 0, t)
|
|
elif model=='ir.actions.wizard':
|
|
service_name = 'wizard.'+obj.wiz_name
|
|
obj2 = netsvc._service[service_name]
|
|
for state_name, state_def in obj2.states.iteritems():
|
|
if 'result' in state_def:
|
|
result = state_def['result']
|
|
if result['type'] != 'form':
|
|
continue
|
|
name = obj.wiz_name + ',' + state_name
|
|
|
|
# export fields
|
|
for field_name, field_def in result['fields'].iteritems():
|
|
if 'string' in field_def:
|
|
source = field_def['string']
|
|
res_name = name + ',' + field_name
|
|
push_translation(module, 'wizard_field', res_name, 0, source)
|
|
|
|
# export arch
|
|
arch = result['arch']
|
|
if not isinstance(arch, UpdateableStr):
|
|
d = xml.dom.minidom.parseString(arch)
|
|
for t in trans_parse_view(d.documentElement):
|
|
push_translation(module, 'wizard_view', name, 0, t)
|
|
|
|
# export button labels
|
|
for but_args in result['state']:
|
|
button_name = but_args[0]
|
|
button_label = but_args[1]
|
|
res_name = name + ',' + button_name
|
|
push_translation(module, 'wizard_button', res_name, 0, button_label)
|
|
|
|
elif model=='ir.model.fields':
|
|
field_name = obj.name
|
|
field_def = pool.get(obj.model)._columns[field_name]
|
|
|
|
name = obj.model + "," + field_name
|
|
push_translation(module, 'field', name, 0, field_def.string.encode('utf8'))
|
|
|
|
if field_def.help:
|
|
push_translation(module, 'help', name, 0, field_def.help.encode('utf8'))
|
|
|
|
if field_def.translate:
|
|
ids = pool.get(obj.model).search(cr, uid, [])
|
|
obj_values = pool.get(obj.model).read(cr, uid, ids, [field_name])
|
|
for obj_value in obj_values:
|
|
res_id = obj_value['id']
|
|
if obj.name in ('ir.model', 'ir.ui.menu'):
|
|
res_id = 0
|
|
model_data_ids = model_data_obj.search(cr, uid, [
|
|
('model', '=', model),
|
|
('res_id', '=', res_id),
|
|
])
|
|
if not model_data_ids:
|
|
push_translation(module, 'model', name, 0, obj_value[field_name])
|
|
|
|
if hasattr(field_def, 'selection') and isinstance(field_def.selection, (list, tuple)):
|
|
for key, val in field_def.selection:
|
|
push_translation(module, 'selection', name, 0, val.encode('utf8'))
|
|
|
|
elif model=='ir.actions.report.xml':
|
|
name = obj.report_name
|
|
fname = ""
|
|
if obj.report_rml:
|
|
fname = obj.report_rml
|
|
parse_func = trans_parse_rml
|
|
report_type = "rml"
|
|
elif obj.report_xsl:
|
|
fname = obj.report_xsl
|
|
parse_func = trans_parse_xsl
|
|
report_type = "xsl"
|
|
try:
|
|
xmlstr = tools.file_open(fname).read()
|
|
d = xml.dom.minidom.parseString(xmlstr)
|
|
for t in parse_func(d.documentElement):
|
|
push_translation(module, report_type, name, 0, t)
|
|
except IOError:
|
|
if fname:
|
|
logger.notifyChannel("init", netsvc.LOG_WARNING, "couldn't export translation for report %s %s %s" % (name, report_type, fname))
|
|
|
|
for constraint in pool.get(model)._constraints:
|
|
msg = constraint[1]
|
|
push_translation(module, 'constraint', model, 0, msg.encode('utf8'))
|
|
|
|
for field_name,field_def in pool.get(model)._columns.items():
|
|
if field_def.translate:
|
|
name = model + "," + field_name
|
|
trad = getattr(obj, field_name) or ''
|
|
push_translation(module, 'model', name, xml_name, trad)
|
|
|
|
# parse source code for _() calls
|
|
def get_module_from_path(path):
|
|
relative_addons_path = tools.config['addons_path'][len(tools.config['root_path'])+1:]
|
|
if path.startswith(relative_addons_path):
|
|
path = path[len(relative_addons_path)+1:]
|
|
return path.split(os.path.sep)[0]
|
|
return 'base' # files that are not in a module are considered as being in 'base' module
|
|
|
|
modobj = pool.get('ir.module.module')
|
|
for root, dirs, files in os.walk(tools.config['root_path']):
|
|
for fname in fnmatch.filter(files, '*.py'):
|
|
fabsolutepath = join(root, fname)
|
|
frelativepath = fabsolutepath[len(tools.config['root_path'])+1:]
|
|
module = get_module_from_path(frelativepath)
|
|
is_mod_installed = modobj.search(cr, uid, [('state', '=', 'installed'), ('name', '=', module)]) <> []
|
|
if (('all' in modules) or (module in modules)) and is_mod_installed:
|
|
code_string = tools.file_open(fabsolutepath, subdir='').read()
|
|
iter = re.finditer(
|
|
'[^a-zA-Z0-9_]_\([\s]*["\'](.+?)["\'][\s]*\)',
|
|
code_string)
|
|
for i in iter:
|
|
push_translation(module, 'code', frelativepath, 0, i.group(1).encode('utf8'))
|
|
|
|
|
|
out = [["module","type","name","res_id","src","value"]] # header
|
|
# translate strings marked as to be translated
|
|
for module, type, name, id, source in _to_translate:
|
|
trans = trans_obj._get_source(cr, uid, name, type, lang, source)
|
|
out.append([module, type, name, id, source, trans or ''])
|
|
|
|
cr.close()
|
|
return out
|
|
|
|
def trans_load(db_name, filename, lang, strict=False):
|
|
logger = netsvc.Logger()
|
|
try:
|
|
fileobj = open(filename,'r')
|
|
fileformat = os.path.splitext(filename)[-1][1:].lower()
|
|
r = trans_load_data(db_name, fileobj, fileformat, lang, strict=False)
|
|
fileobj.close()
|
|
return r
|
|
except IOError:
|
|
logger.notifyChannel("init", netsvc.LOG_ERROR, "couldn't read file")
|
|
return None
|
|
|
|
def trans_load_data(db_name, fileobj, fileformat, lang, strict=False, lang_name=None):
|
|
logger = netsvc.Logger()
|
|
logger.notifyChannel("init", netsvc.LOG_INFO,
|
|
'loading translation file for language %s' % (lang))
|
|
pool = pooler.get_pool(db_name)
|
|
lang_obj = pool.get('res.lang')
|
|
trans_obj = pool.get('ir.translation')
|
|
model_data_obj = pool.get('ir.model.data')
|
|
try:
|
|
uid = 1
|
|
cr = pooler.get_db(db_name).cursor()
|
|
|
|
ids = lang_obj.search(cr, uid, [('code','=',lang)])
|
|
if not ids:
|
|
if not lang_name:
|
|
lang_name=lang
|
|
languages=tools.get_languages()
|
|
if lang in languages:
|
|
lang_name=languages[lang]
|
|
ids = lang_obj.create(cr, uid, {
|
|
'code': lang,
|
|
'name': lang_name,
|
|
'translatable': 1,
|
|
})
|
|
else:
|
|
lang_obj.write(cr, uid, ids, {'translatable':1})
|
|
lang_ids = lang_obj.search(cr, uid, [])
|
|
langs = lang_obj.read(cr, uid, lang_ids)
|
|
ls = map(lambda x: (x['code'],x['name']), langs)
|
|
|
|
fileobj.seek(0)
|
|
|
|
if fileformat == 'csv':
|
|
reader = csv.reader(fileobj, quotechar='"', delimiter=',')
|
|
# read the first line of the file (it contains columns titles)
|
|
for row in reader:
|
|
f = row
|
|
break
|
|
elif fileformat == 'po':
|
|
reader = TinyPoFile(fileobj)
|
|
f = ['type', 'name', 'res_id', 'src', 'value']
|
|
else:
|
|
raise Exception(_('Bad file format'))
|
|
|
|
# read the rest of the file
|
|
line = 1
|
|
for row in reader:
|
|
line += 1
|
|
#try:
|
|
# skip empty rows and rows where the translation field (=last fiefd) is empty
|
|
if (not row) or (not row[-1]):
|
|
#print "translate: skip %s" % repr(row)
|
|
continue
|
|
|
|
# dictionary which holds values for this line of the csv file
|
|
# {'lang': ..., 'type': ..., 'name': ..., 'res_id': ...,
|
|
# 'src': ..., 'value': ...}
|
|
dic = {'lang': lang}
|
|
for i in range(len(f)):
|
|
if f[i] in ('module',):
|
|
continue
|
|
dic[f[i]] = row[i]
|
|
|
|
try:
|
|
dic['res_id'] = int(dic['res_id'])
|
|
except:
|
|
model_data_ids = model_data_obj.search(cr, uid, [
|
|
('model', '=', dic['name'].split(',')[0]),
|
|
('module', '=', dic['res_id'].split('.', 1)[0]),
|
|
('name', '=', dic['res_id'].split('.', 1)[1]),
|
|
])
|
|
if model_data_ids:
|
|
dic['res_id'] = model_data_obj.browse(cr, uid,
|
|
model_data_ids[0]).res_id
|
|
else:
|
|
dic['res_id'] = False
|
|
|
|
if dic['type'] == 'model' and not strict:
|
|
(model, field) = dic['name'].split(',')
|
|
|
|
# get the ids of the resources of this model which share
|
|
# the same source
|
|
obj = pool.get(model)
|
|
if obj:
|
|
ids = obj.search(cr, uid, [(field, '=', dic['src'])])
|
|
|
|
# if the resource id (res_id) is in that list, use it,
|
|
# otherwise use the whole list
|
|
ids = (dic['res_id'] in ids) and [dic['res_id']] or ids
|
|
for id in ids:
|
|
dic['res_id'] = id
|
|
ids = trans_obj.search(cr, uid, [
|
|
('lang', '=', lang),
|
|
('type', '=', dic['type']),
|
|
('name', '=', dic['name']),
|
|
('src', '=', dic['src']),
|
|
('res_id', '=', dic['res_id'])
|
|
])
|
|
if ids:
|
|
trans_obj.write(cr, uid, ids, {'value': dic['value']})
|
|
else:
|
|
trans_obj.create(cr, uid, dic)
|
|
else:
|
|
ids = trans_obj.search(cr, uid, [
|
|
('lang', '=', lang),
|
|
('type', '=', dic['type']),
|
|
('name', '=', dic['name']),
|
|
('src', '=', dic['src'])
|
|
])
|
|
if ids:
|
|
trans_obj.write(cr, uid, ids, {'value': dic['value']})
|
|
else:
|
|
trans_obj.create(cr, uid, dic)
|
|
cr.commit()
|
|
#except Exception, e:
|
|
# logger.notifyChannel('init', netsvc.LOG_ERROR,
|
|
# 'Import error: %s on line %d: %s!' % (str(e), line, row))
|
|
# cr.rollback()
|
|
# cr.close()
|
|
# cr = pooler.get_db(db_name).cursor()
|
|
cr.close()
|
|
logger.notifyChannel("init", netsvc.LOG_INFO,
|
|
"translation file loaded succesfully")
|
|
except IOError:
|
|
logger.notifyChannel("init", netsvc.LOG_ERROR, "couldn't read file")
|
|
|