planetiler/planetiler-core/src/main/java/com/onthegomap/planetiler/reader/NaturalEarthReader.java

188 wiersze
6.7 KiB
Java

package com.onthegomap.planetiler.reader;
import com.onthegomap.planetiler.Profile;
import com.onthegomap.planetiler.collection.FeatureGroup;
import com.onthegomap.planetiler.config.PlanetilerConfig;
import com.onthegomap.planetiler.geo.GeoUtils;
import com.onthegomap.planetiler.stats.Stats;
import com.onthegomap.planetiler.util.FileUtils;
import com.onthegomap.planetiler.util.LogUtil;
import com.onthegomap.planetiler.worker.WorkerPipeline;
import java.io.IOException;
import java.nio.file.FileSystems;
import java.nio.file.Files;
import java.nio.file.Path;
import java.nio.file.StandardCopyOption;
import java.sql.Connection;
import java.sql.DriverManager;
import java.sql.ResultSet;
import java.sql.SQLException;
import java.sql.Statement;
import java.util.ArrayList;
import java.util.HashMap;
import java.util.List;
import org.locationtech.jts.geom.Geometry;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
/**
* Utility that reads {@link SourceFeature SourceFeatures} from the geometries contained in a Natural Earth sqlite
* distribution.
*
* @see <a href="https://www.naturalearthdata.com/">Natural Earth</a>
*/
public class NaturalEarthReader extends SimpleReader {
private static final Logger LOGGER = LoggerFactory.getLogger(NaturalEarthReader.class);
private final Connection conn;
private Path extracted;
static {
// make sure sqlite driver loaded
try {
Class.forName("org.sqlite.JDBC");
} catch (ClassNotFoundException e) {
throw new IllegalStateException("sqlite JDBC driver not found");
}
}
NaturalEarthReader(String sourceName, Path input, Path tmpDir, Profile profile, Stats stats) {
super(profile, stats, sourceName);
LogUtil.setStage(sourceName);
try {
conn = open(input, tmpDir);
} catch (IOException | SQLException e) {
throw new IllegalArgumentException(e);
}
}
/**
* Renders map features for all elements from a Natural Earth sqlite file, or zip file containing a sqlite file, based
* on the mapping logic defined in {@code profile}.
*
* @param sourceName string ID for this reader to use in logs and stats
* @param input path to the sqlite or zip file
* @param tmpDir directory to extract the sqlite file into (if input is a zip file)
* @param writer consumer for rendered features
* @param config user-defined parameters controlling number of threads and log interval
* @param profile logic that defines what map features to emit for each source feature
* @param stats to keep track of counters and timings
* @throws IllegalArgumentException if a problem occurs reading the input file
*/
public static void process(String sourceName, Path input, Path tmpDir, FeatureGroup writer, PlanetilerConfig config,
Profile profile, Stats stats) {
try (var reader = new NaturalEarthReader(sourceName, input, tmpDir, profile, stats)) {
reader.process(writer, config);
}
}
/** Returns a JDBC connection to the sqlite file. Input can be the sqlite file itself or a zip file containing it. */
private Connection open(Path path, Path tmpLocation) throws IOException, SQLException {
String uri = "jdbc:sqlite:" + path.toAbsolutePath();
if (FileUtils.hasExtension(path, "zip")) {
Path toOpen = tmpLocation == null ? Files.createTempFile("sqlite", "natearth") : tmpLocation;
extracted = toOpen;
try (var zipFs = FileSystems.newFileSystem(path)) {
var zipEntry = FileUtils.walkFileSystem(zipFs)
.filter(Files::isRegularFile)
.filter(entry -> FileUtils.hasExtension(entry, "sqlite"))
.findFirst()
.orElseThrow(() -> new IllegalArgumentException("No .sqlite file found inside " + path));
LOGGER.info("unzipping " + path.toAbsolutePath() + " to " + extracted);
Files.copy(Files.newInputStream(zipEntry), extracted, StandardCopyOption.REPLACE_EXISTING);
extracted.toFile().deleteOnExit();
}
uri = "jdbc:sqlite:" + toOpen.toAbsolutePath();
}
return DriverManager.getConnection(uri);
}
private List<String> tableNames() {
List<String> result = new ArrayList<>();
try (ResultSet rs = conn.getMetaData().getTables(null, null, null, null)) {
while (rs.next()) {
String table = rs.getString("TABLE_NAME");
result.add(table);
}
} catch (SQLException e) {
throw new RuntimeException(e);
}
return result;
}
@Override
public long getCount() {
long count = 0;
for (String table : tableNames()) {
try (
var stmt = conn.createStatement();
var result = stmt.executeQuery("select count(*) from " + table + " where GEOMETRY is not null;")
) {
count += result.getLong(1);
} catch (SQLException e) {
// maybe no GEOMETRY column?
}
}
return count;
}
@Override
public WorkerPipeline.SourceStep<SimpleFeature> read() {
return next -> {
long id = 0;
// pass every element in every table through the profile
var tables = tableNames();
for (int i = 0; i < tables.size(); i++) {
String table = tables.get(i);
LOGGER.trace("Naturalearth loading " + i + "/" + tables.size() + ": " + table);
try (Statement statement = conn.createStatement()) {
ResultSet rs = statement.executeQuery("select * from " + table + ";");
String[] column = new String[rs.getMetaData().getColumnCount()];
int geometryColumn = -1;
for (int c = 0; c < column.length; c++) {
String name = rs.getMetaData().getColumnName(c + 1);
column[c] = name;
if ("GEOMETRY".equals(name)) {
geometryColumn = c;
}
}
if (geometryColumn >= 0) {
while (rs.next()) {
byte[] geometry = rs.getBytes(geometryColumn + 1);
if (geometry == null) {
continue;
}
// create the feature and pass to next stage
Geometry latLonGeometry = GeoUtils.WKB_READER.read(geometry);
SimpleFeature readerGeometry = SimpleFeature.create(latLonGeometry, new HashMap<>(column.length - 1),
sourceName, table, id);
for (int c = 0; c < column.length; c++) {
if (c != geometryColumn) {
Object value = rs.getObject(c + 1);
String key = column[c];
readerGeometry.setTag(key, value);
}
}
next.accept(readerGeometry);
}
}
}
}
};
}
@Override
public void close() {
try {
conn.close();
} catch (SQLException e) {
LOGGER.error("Error closing sqlite file", e);
}
if (extracted != null) {
FileUtils.deleteFile(extracted);
}
}
}