mirror of
https://github.com/protomaps/PMTiles.git
synced 2026-02-04 19:01:08 +00:00
Merge pull request #22 from eddy-geek/ed
Refactor + pmtiles to mbtiles fix
This commit is contained in:
@@ -2,13 +2,10 @@
|
|||||||
|
|
||||||
#pmtiles to files
|
#pmtiles to files
|
||||||
import argparse
|
import argparse
|
||||||
import gzip
|
|
||||||
import json
|
|
||||||
import os
|
import os
|
||||||
import sqlite3
|
|
||||||
import shutil
|
import shutil
|
||||||
from pmtiles.reader import read
|
|
||||||
from pmtiles.writer import write
|
from pmtiles.convert import mbtiles_to_pmtiles, pmtiles_to_mbtiles, pmtiles_to_dir
|
||||||
|
|
||||||
parser = argparse.ArgumentParser(description='Convert between PMTiles and other archive formats.')
|
parser = argparse.ArgumentParser(description='Convert between PMTiles and other archive formats.')
|
||||||
parser.add_argument('input',help='Input .mbtiles or .pmtiles')
|
parser.add_argument('input',help='Input .mbtiles or .pmtiles')
|
||||||
@@ -18,11 +15,6 @@ parser.add_argument('--gzip', help='Add gzip encoding to the output if it is not
|
|||||||
parser.add_argument('--overwrite', help='Overwrite the existing output.',action='store_true')
|
parser.add_argument('--overwrite', help='Overwrite the existing output.',action='store_true')
|
||||||
args = parser.parse_args()
|
args = parser.parse_args()
|
||||||
|
|
||||||
def may_compress(data,compress):
|
|
||||||
if compress and data[0:2] != b'\x1f\x8b':
|
|
||||||
return gzip.compress(data)
|
|
||||||
return data
|
|
||||||
|
|
||||||
if os.path.exists(args.output) and not args.overwrite:
|
if os.path.exists(args.output) and not args.overwrite:
|
||||||
print("Output exists, use --overwrite to overwrite the output.")
|
print("Output exists, use --overwrite to overwrite the output.")
|
||||||
exit(1)
|
exit(1)
|
||||||
@@ -33,55 +25,13 @@ if args.overwrite:
|
|||||||
shutil.rmtree(args.output)
|
shutil.rmtree(args.output)
|
||||||
|
|
||||||
if args.input.endswith('.mbtiles') and args.output.endswith('.pmtiles'):
|
if args.input.endswith('.mbtiles') and args.output.endswith('.pmtiles'):
|
||||||
conn = sqlite3.connect(args.input)
|
mbtiles_to_pmtiles(args.input, args.output, args.maxzoom)
|
||||||
cursor = conn.cursor()
|
|
||||||
|
|
||||||
with write(args.output) as writer:
|
|
||||||
for row in cursor.execute('SELECT zoom_level,tile_column,tile_row,tile_data FROM tiles WHERE zoom_level <= ? ORDER BY zoom_level,tile_column,tile_row ASC',(args.maxzoom or 99,)):
|
|
||||||
flipped = (1 << row[0]) - 1 - row[2]
|
|
||||||
writer.write_tile(row[0],row[1],flipped,row[3])
|
|
||||||
|
|
||||||
metadata = {}
|
|
||||||
for row in cursor.execute('SELECT name,value FROM metadata'):
|
|
||||||
metadata[row[0]] = row[1]
|
|
||||||
if args.maxzoom:
|
|
||||||
metadata['maxzoom'] = str(args.maxzoom)
|
|
||||||
result = writer.finalize(metadata)
|
|
||||||
print("Num tiles:",result['num_tiles'])
|
|
||||||
print("Num unique tiles:",result['num_unique_tiles'])
|
|
||||||
print("Num leaves:",result['num_leaves'])
|
|
||||||
|
|
||||||
conn.close()
|
|
||||||
elif args.input.endswith('.pmtiles') and args.output.endswith('.mbtiles'):
|
elif args.input.endswith('.pmtiles') and args.output.endswith('.mbtiles'):
|
||||||
conn = sqlite3.connect(args.output)
|
pmtiles_to_mbtiles(args.input, args.output, args.gzip)
|
||||||
cursor = conn.cursor()
|
|
||||||
cursor.execute('CREATE TABLE metadata (name text, value text);')
|
|
||||||
cursor.execute('CREATE TABLE tiles (zoom_level integer, tile_column integer, tile_row integer, tile_data blob);')
|
|
||||||
|
|
||||||
with read(args.input) as reader:
|
|
||||||
for k,v in reader.metadata.items():
|
|
||||||
cursor.execute('INSERT INTO metadata VALUES(?,?)',(k,v))
|
|
||||||
for tile, data in reader.tiles():
|
|
||||||
cursor.execute('INSERT INTO tiles VALUES(?,?,?,?)',(tile[0],tile[1],tile[2],may_compress(data,args.gzip)))
|
|
||||||
|
|
||||||
cursor.execute('CREATE UNIQUE INDEX tile_index on tiles (zoom_level, tile_column, tile_row);')
|
|
||||||
conn.commit()
|
|
||||||
conn.close()
|
|
||||||
elif args.input.endswith(".pmtiles"):
|
elif args.input.endswith(".pmtiles"):
|
||||||
os.makedirs(args.output)
|
pmtiles_to_dir(args.input, args.output, args.gzip)
|
||||||
|
|
||||||
with read(args.input) as reader:
|
|
||||||
metadata = reader.metadata
|
|
||||||
metadata['format']
|
|
||||||
with open(os.path.join(args.output,'metadata.json'),'w') as f:
|
|
||||||
f.write(json.dumps(metadata))
|
|
||||||
|
|
||||||
for tile, data in reader.tiles():
|
|
||||||
directory = os.path.join(args.output,str(tile[0]),str(tile[1]))
|
|
||||||
path = os.path.join(directory,str(tile[2]) + '.' + metadata['format'])
|
|
||||||
os.makedirs(directory,exist_ok=True)
|
|
||||||
with open(path,'wb') as f:
|
|
||||||
f.write(may_compress(data,args.gzip))
|
|
||||||
|
|
||||||
else:
|
else:
|
||||||
print("Conversion not implemented")
|
print("Conversion not implemented")
|
||||||
|
|||||||
70
python/pmtiles/convert.py
Normal file
70
python/pmtiles/convert.py
Normal file
@@ -0,0 +1,70 @@
|
|||||||
|
|
||||||
|
#pmtiles to files
|
||||||
|
import gzip
|
||||||
|
import json
|
||||||
|
import os
|
||||||
|
import sqlite3
|
||||||
|
from pmtiles.reader import read
|
||||||
|
from pmtiles.writer import write
|
||||||
|
|
||||||
|
|
||||||
|
def may_compress(data,compress):
|
||||||
|
if compress and data[0:2] != b'\x1f\x8b':
|
||||||
|
return gzip.compress(data)
|
||||||
|
return data
|
||||||
|
|
||||||
|
|
||||||
|
def mbtiles_to_pmtiles(input, output, maxzoom):
|
||||||
|
conn = sqlite3.connect(input)
|
||||||
|
cursor = conn.cursor()
|
||||||
|
|
||||||
|
with write(output) as writer:
|
||||||
|
for row in cursor.execute('SELECT zoom_level,tile_column,tile_row,tile_data FROM tiles WHERE zoom_level <= ? ORDER BY zoom_level,tile_column,tile_row ASC',(maxzoom or 99,)):
|
||||||
|
flipped = (1 << row[0]) - 1 - row[2]
|
||||||
|
writer.write_tile(row[0],row[1],flipped,row[3])
|
||||||
|
|
||||||
|
metadata = {}
|
||||||
|
for row in cursor.execute('SELECT name,value FROM metadata'):
|
||||||
|
metadata[row[0]] = row[1]
|
||||||
|
if maxzoom:
|
||||||
|
metadata['maxzoom'] = str(maxzoom)
|
||||||
|
result = writer.finalize(metadata)
|
||||||
|
print("Num tiles:",result['num_tiles'])
|
||||||
|
print("Num unique tiles:",result['num_unique_tiles'])
|
||||||
|
print("Num leaves:",result['num_leaves'])
|
||||||
|
|
||||||
|
conn.close()
|
||||||
|
|
||||||
|
|
||||||
|
def pmtiles_to_mbtiles(input, output, gzip):
|
||||||
|
conn = sqlite3.connect(output)
|
||||||
|
cursor = conn.cursor()
|
||||||
|
cursor.execute('CREATE TABLE metadata (name text, value text);')
|
||||||
|
cursor.execute('CREATE TABLE tiles (zoom_level integer, tile_column integer, tile_row integer, tile_data blob);')
|
||||||
|
|
||||||
|
with read(input) as reader:
|
||||||
|
for k,v in reader.metadata.items():
|
||||||
|
cursor.execute('INSERT INTO metadata VALUES(?,?)',(k,v))
|
||||||
|
for tile, data in reader.tiles():
|
||||||
|
flipped = (1 << tile[0]) - 1 - tile[2]
|
||||||
|
cursor.execute('INSERT INTO tiles VALUES(?,?,?,?)',(tile[0],tile[1],flipped,may_compress(data,gzip)))
|
||||||
|
|
||||||
|
cursor.execute('CREATE UNIQUE INDEX tile_index on tiles (zoom_level, tile_column, tile_row);')
|
||||||
|
conn.commit()
|
||||||
|
conn.close()
|
||||||
|
|
||||||
|
def pmtiles_to_dir(input, output, gzip):
|
||||||
|
os.makedirs(output)
|
||||||
|
|
||||||
|
with read(input) as reader:
|
||||||
|
metadata = reader.metadata
|
||||||
|
metadata['format']
|
||||||
|
with open(os.path.join(output,'metadata.json'),'w') as f:
|
||||||
|
f.write(json.dumps(metadata))
|
||||||
|
|
||||||
|
for tile, data in reader.tiles():
|
||||||
|
directory = os.path.join(output,str(tile[0]),str(tile[1]))
|
||||||
|
path = os.path.join(directory,str(tile[2]) + '.' + metadata['format'])
|
||||||
|
os.makedirs(directory,exist_ok=True)
|
||||||
|
with open(path,'wb') as f:
|
||||||
|
f.write(may_compress(data,gzip))
|
||||||
Reference in New Issue
Block a user