Use LUTs in SpineTocCache
This commit is contained in:
parent
05fce6b818
commit
63f0acd852
@ -181,7 +181,7 @@ bool Epub::load() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Now compute mappings and sizes (this loads entries temporarily, computes, then rewrites)
|
// Now compute mappings and sizes (this loads entries temporarily, computes, then rewrites)
|
||||||
if (!spineTocCache->updateMappingsAndSizes(filepath)) {
|
if (!spineTocCache->updateMapsAndSizes(filepath)) {
|
||||||
Serial.printf("[%lu] [EBP] Could not update mappings and sizes\n", millis());
|
Serial.printf("[%lu] [EBP] Could not update mappings and sizes\n", millis());
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|||||||
@ -4,6 +4,24 @@
|
|||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
|
|
||||||
|
bool FsHelpers::openFileForRead(const char* moduleName, const std::string& path, File& file) {
|
||||||
|
file = SD.open(path.c_str(), FILE_READ);
|
||||||
|
if (!file) {
|
||||||
|
Serial.printf("[%lu] [%s] Failed to open file for reading: %s\n", millis(), moduleName, path.c_str());
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
bool FsHelpers::openFileForWrite(const char* moduleName, const std::string& path, File& file) {
|
||||||
|
file = SD.open(path.c_str(), FILE_WRITE, true);
|
||||||
|
if (!file) {
|
||||||
|
Serial.printf("[%lu] [%s] Failed to open spine file for writing: %s\n", millis(), moduleName, path.c_str());
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
bool FsHelpers::removeDir(const char* path) {
|
bool FsHelpers::removeDir(const char* path) {
|
||||||
// 1. Open the directory
|
// 1. Open the directory
|
||||||
File dir = SD.open(path);
|
File dir = SD.open(path);
|
||||||
|
|||||||
@ -1,8 +1,12 @@
|
|||||||
#pragma once
|
#pragma once
|
||||||
|
#include <FS.h>
|
||||||
|
|
||||||
#include <string>
|
#include <string>
|
||||||
|
|
||||||
class FsHelpers {
|
class FsHelpers {
|
||||||
public:
|
public:
|
||||||
|
static bool openFileForRead(const char* moduleName, const std::string& path, File& file);
|
||||||
|
static bool openFileForWrite(const char* moduleName, const std::string& path, File& file);
|
||||||
static bool removeDir(const char* path);
|
static bool removeDir(const char* path);
|
||||||
static std::string normalisePath(const std::string& path);
|
static std::string normalisePath(const std::string& path);
|
||||||
};
|
};
|
||||||
|
|||||||
@ -11,27 +11,10 @@
|
|||||||
|
|
||||||
namespace {
|
namespace {
|
||||||
constexpr uint8_t SPINE_TOC_CACHE_VERSION = 1;
|
constexpr uint8_t SPINE_TOC_CACHE_VERSION = 1;
|
||||||
|
constexpr size_t SPINE_TOC_META_HEADER_SIZE = sizeof(SPINE_TOC_CACHE_VERSION) + sizeof(uint16_t) * 2;
|
||||||
constexpr char spineTocMetaBinFile[] = "/spine_toc_meta.bin";
|
constexpr char spineTocMetaBinFile[] = "/spine_toc_meta.bin";
|
||||||
constexpr char spineBinFile[] = "/spine.bin";
|
constexpr char spineBinFile[] = "/spine.bin";
|
||||||
constexpr char tocBinFile[] = "/toc.bin";
|
constexpr char tocBinFile[] = "/toc.bin";
|
||||||
|
|
||||||
bool openFileForRead(const std::string& path, File& file) {
|
|
||||||
file = SD.open(path.c_str(), FILE_READ);
|
|
||||||
if (!file) {
|
|
||||||
Serial.printf("[%lu] [STC] Failed to open file for reading: %s\n", millis(), path.c_str());
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
|
|
||||||
bool openFileForWrite(const std::string& path, File& file) {
|
|
||||||
file = SD.open(path.c_str(), FILE_WRITE, true);
|
|
||||||
if (!file) {
|
|
||||||
Serial.printf("[%lu] [STC] Failed to open spine file for writing: %s\n", millis(), path.c_str());
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
return true;
|
|
||||||
}
|
|
||||||
} // namespace
|
} // namespace
|
||||||
|
|
||||||
bool SpineTocCache::beginWrite() {
|
bool SpineTocCache::beginWrite() {
|
||||||
@ -42,54 +25,75 @@ bool SpineTocCache::beginWrite() {
|
|||||||
Serial.printf("[%lu] [STC] Beginning write to cache path: %s\n", millis(), cachePath.c_str());
|
Serial.printf("[%lu] [STC] Beginning write to cache path: %s\n", millis(), cachePath.c_str());
|
||||||
|
|
||||||
// Open spine file for writing
|
// Open spine file for writing
|
||||||
if (!openFileForWrite(cachePath + spineBinFile, spineFile)) {
|
if (!FsHelpers::openFileForWrite("STC", cachePath + spineBinFile, spineFile)) {
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Open TOC file for writing
|
// Open TOC file for writing
|
||||||
if (!openFileForWrite(cachePath + tocBinFile, tocFile)) {
|
if (!FsHelpers::openFileForWrite("STC", cachePath + tocBinFile, tocFile)) {
|
||||||
spineFile.close();
|
spineFile.close();
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Open meta file for writing
|
||||||
|
if (!FsHelpers::openFileForWrite("STC", cachePath + spineTocMetaBinFile, metaFile)) {
|
||||||
|
spineFile.close();
|
||||||
|
tocFile.close();
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Write 0s into first slots, LUT is written during `addSpineEntry` and `addTocEntry`, and counts are rewritten at
|
||||||
|
// the end
|
||||||
|
serialization::writePod(metaFile, SPINE_TOC_CACHE_VERSION);
|
||||||
|
serialization::writePod(metaFile, spineCount);
|
||||||
|
serialization::writePod(metaFile, tocCount);
|
||||||
|
|
||||||
Serial.printf("[%lu] [STC] Began writing cache files\n", millis());
|
Serial.printf("[%lu] [STC] Began writing cache files\n", millis());
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
void SpineTocCache::writeSpineEntry(File& file, const SpineEntry& entry) const {
|
size_t SpineTocCache::writeSpineEntry(File& file, const SpineEntry& entry) const {
|
||||||
|
const auto pos = file.position();
|
||||||
serialization::writeString(file, entry.href);
|
serialization::writeString(file, entry.href);
|
||||||
serialization::writePod(file, entry.cumulativeSize);
|
serialization::writePod(file, entry.cumulativeSize);
|
||||||
serialization::writePod(file, entry.tocIndex);
|
serialization::writePod(file, entry.tocIndex);
|
||||||
|
return pos;
|
||||||
}
|
}
|
||||||
|
|
||||||
void SpineTocCache::writeTocEntry(File& file, const TocEntry& entry) const {
|
size_t SpineTocCache::writeTocEntry(File& file, const TocEntry& entry) const {
|
||||||
|
const auto pos = file.position();
|
||||||
serialization::writeString(file, entry.title);
|
serialization::writeString(file, entry.title);
|
||||||
serialization::writeString(file, entry.href);
|
serialization::writeString(file, entry.href);
|
||||||
serialization::writeString(file, entry.anchor);
|
serialization::writeString(file, entry.anchor);
|
||||||
serialization::writePod(file, entry.level);
|
serialization::writePod(file, entry.level);
|
||||||
serialization::writePod(file, entry.spineIndex);
|
serialization::writePod(file, entry.spineIndex);
|
||||||
|
return pos;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Note: for the LUT to be accurate, this **MUST** be called for all spine items before `addTocEntry` is ever called
|
||||||
|
// this is because in this function we're marking positions of the items
|
||||||
void SpineTocCache::addSpineEntry(const std::string& href) {
|
void SpineTocCache::addSpineEntry(const std::string& href) {
|
||||||
if (!buildMode || !spineFile) {
|
if (!buildMode || !spineFile || !metaFile) {
|
||||||
Serial.printf("[%lu] [STC] addSpineEntry called but not in build mode\n", millis());
|
Serial.printf("[%lu] [STC] addSpineEntry called but not in build mode\n", millis());
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
const SpineEntry entry(href, 0, -1);
|
const SpineEntry entry(href, 0, -1);
|
||||||
writeSpineEntry(spineFile, entry);
|
const auto position = writeSpineEntry(spineFile, entry);
|
||||||
|
serialization::writePod(metaFile, position);
|
||||||
spineCount++;
|
spineCount++;
|
||||||
}
|
}
|
||||||
|
|
||||||
void SpineTocCache::addTocEntry(const std::string& title, const std::string& href, const std::string& anchor,
|
void SpineTocCache::addTocEntry(const std::string& title, const std::string& href, const std::string& anchor,
|
||||||
const uint8_t level) {
|
const uint8_t level) {
|
||||||
if (!buildMode || !tocFile) {
|
if (!buildMode || !tocFile || !metaFile) {
|
||||||
Serial.printf("[%lu] [STC] addTocEntry called but not in build mode\n", millis());
|
Serial.printf("[%lu] [STC] addTocEntry called but not in build mode\n", millis());
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
const TocEntry entry(title, href, anchor, level, -1);
|
const TocEntry entry(title, href, anchor, level, -1);
|
||||||
writeTocEntry(tocFile, entry);
|
const auto position = writeTocEntry(tocFile, entry);
|
||||||
|
serialization::writePod(metaFile, position);
|
||||||
tocCount++;
|
tocCount++;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -102,12 +106,8 @@ bool SpineTocCache::endWrite() {
|
|||||||
spineFile.close();
|
spineFile.close();
|
||||||
tocFile.close();
|
tocFile.close();
|
||||||
|
|
||||||
// Write metadata files with counts
|
// Write correct counts into meta file
|
||||||
File metaFile;
|
metaFile.seek(sizeof(SPINE_TOC_CACHE_VERSION));
|
||||||
if (!openFileForWrite(cachePath + spineTocMetaBinFile, metaFile)) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
serialization::writePod(metaFile, SPINE_TOC_CACHE_VERSION);
|
|
||||||
serialization::writePod(metaFile, spineCount);
|
serialization::writePod(metaFile, spineCount);
|
||||||
serialization::writePod(metaFile, tocCount);
|
serialization::writePod(metaFile, tocCount);
|
||||||
metaFile.close();
|
metaFile.close();
|
||||||
@ -135,21 +135,16 @@ SpineTocCache::TocEntry SpineTocCache::readTocEntry(File& file) const {
|
|||||||
return entry;
|
return entry;
|
||||||
}
|
}
|
||||||
|
|
||||||
bool SpineTocCache::updateMappingsAndSizes(const std::string& epubPath) {
|
bool SpineTocCache::updateMapsAndSizes(const std::string& epubPath) {
|
||||||
Serial.printf("[%lu] [STC] Computing mappings and sizes for %d spine, %d TOC entries\n", millis(), spineCount,
|
Serial.printf("[%lu] [STC] Computing mappings and sizes for %d spine, %d TOC entries\n", millis(), spineCount,
|
||||||
tocCount);
|
tocCount);
|
||||||
|
|
||||||
// Read all spine and TOC entries into temporary arrays (we need them all to compute mappings)
|
|
||||||
// TODO: can we do this a bit smarter and avoid loading everything?
|
|
||||||
std::vector<SpineEntry> spineEntries;
|
std::vector<SpineEntry> spineEntries;
|
||||||
std::vector<TocEntry> tocEntries;
|
|
||||||
|
|
||||||
spineEntries.reserve(spineCount);
|
spineEntries.reserve(spineCount);
|
||||||
tocEntries.reserve(tocCount);
|
|
||||||
|
|
||||||
// Read spine entries
|
// Load only the spine items, update them in memory while loading one TOC at a time and storing it
|
||||||
{
|
{
|
||||||
if (!openFileForRead(cachePath + spineBinFile, spineFile)) {
|
if (!FsHelpers::openFileForRead("STC", cachePath + spineBinFile, spineFile)) {
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
for (int i = 0; i < spineCount; i++) {
|
for (int i = 0; i < spineCount; i++) {
|
||||||
@ -158,48 +153,66 @@ bool SpineTocCache::updateMappingsAndSizes(const std::string& epubPath) {
|
|||||||
spineFile.close();
|
spineFile.close();
|
||||||
}
|
}
|
||||||
|
|
||||||
// Read TOC entries
|
// Iterate over TOC entries and update them with the spine mapping
|
||||||
|
// We do this by moving the TOC file and then making a new one parsing through both at the same time
|
||||||
{
|
{
|
||||||
if (!openFileForRead(cachePath + tocBinFile, tocFile)) {
|
SD.rename((cachePath + tocBinFile).c_str(), (cachePath + tocBinFile + ".tmp").c_str());
|
||||||
|
File tempTocFile;
|
||||||
|
if (!FsHelpers::openFileForRead("STC", cachePath + tocBinFile + ".tmp", tempTocFile)) {
|
||||||
|
SD.remove((cachePath + tocBinFile + ".tmp").c_str());
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
if (!FsHelpers::openFileForWrite("STC", cachePath + tocBinFile, tocFile)) {
|
||||||
|
tempTocFile.close();
|
||||||
|
SD.remove((cachePath + tocBinFile + ".tmp").c_str());
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
for (int i = 0; i < tocCount; i++) {
|
for (int i = 0; i < tocCount; i++) {
|
||||||
tocEntries.push_back(readTocEntry(tocFile));
|
auto tocEntry = readTocEntry(tempTocFile);
|
||||||
|
|
||||||
|
// Find the matching spine entry
|
||||||
|
for (int j = 0; j < spineCount; j++) {
|
||||||
|
if (spineEntries[j].href == tocEntry.href) {
|
||||||
|
tocEntry.spineIndex = static_cast<int16_t>(j);
|
||||||
|
// Point the spine to the first TOC entry we come across (in the case that there are multiple)
|
||||||
|
if (spineEntries[j].tocIndex == -1) spineEntries[j].tocIndex = static_cast<int16_t>(i);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
writeTocEntry(tocFile, tocEntry);
|
||||||
}
|
}
|
||||||
tocFile.close();
|
tocFile.close();
|
||||||
|
tempTocFile.close();
|
||||||
|
SD.remove((cachePath + tocBinFile + ".tmp").c_str());
|
||||||
}
|
}
|
||||||
|
|
||||||
// Compute cumulative sizes
|
// By this point all the spine items in memory should have the right `tocIndex` and the TOC file is complete
|
||||||
const ZipFile zip("/sd" + epubPath);
|
|
||||||
size_t cumSize = 0;
|
|
||||||
|
|
||||||
for (int i = 0; i < spineCount; i++) {
|
// Next, compute cumulative sizes
|
||||||
size_t itemSize = 0;
|
{
|
||||||
const std::string path = FsHelpers::normalisePath(spineEntries[i].href);
|
const ZipFile zip("/sd" + epubPath);
|
||||||
if (zip.getInflatedFileSize(path.c_str(), &itemSize)) {
|
size_t cumSize = 0;
|
||||||
cumSize += itemSize;
|
|
||||||
spineEntries[i].cumulativeSize = cumSize;
|
|
||||||
} else {
|
|
||||||
Serial.printf("[%lu] [STC] Warning: Could not get size for spine item: %s\n", millis(), path.c_str());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
Serial.printf("[%lu] [STC] Book size: %lu\n", millis(), cumSize);
|
for (int i = 0; i < spineCount; i++) {
|
||||||
|
size_t itemSize = 0;
|
||||||
// Compute spine <-> TOC mappings
|
const std::string path = FsHelpers::normalisePath(spineEntries[i].href);
|
||||||
for (int i = 0; i < spineCount; i++) {
|
if (zip.getInflatedFileSize(path.c_str(), &itemSize)) {
|
||||||
for (int j = 0; j < tocCount; j++) {
|
cumSize += itemSize;
|
||||||
if (tocEntries[j].href == spineEntries[i].href) {
|
spineEntries[i].cumulativeSize = cumSize;
|
||||||
spineEntries[i].tocIndex = static_cast<int16_t>(j);
|
} else {
|
||||||
tocEntries[j].spineIndex = static_cast<int16_t>(i);
|
Serial.printf("[%lu] [STC] Warning: Could not get size for spine item: %s\n", millis(), path.c_str());
|
||||||
break;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Serial.printf("[%lu] [STC] Book size: %lu\n", millis(), cumSize);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Rewrite spine file with updated data
|
// Rewrite spine file with updated data
|
||||||
{
|
{
|
||||||
if (!openFileForWrite(cachePath + spineBinFile, spineFile)) {
|
if (!FsHelpers::openFileForWrite("STC", cachePath + spineBinFile, spineFile)) {
|
||||||
|
// metaFile.close();
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
for (const auto& entry : spineEntries) {
|
for (const auto& entry : spineEntries) {
|
||||||
@ -208,31 +221,18 @@ bool SpineTocCache::updateMappingsAndSizes(const std::string& epubPath) {
|
|||||||
spineFile.close();
|
spineFile.close();
|
||||||
}
|
}
|
||||||
|
|
||||||
// Rewrite TOC file with updated data
|
|
||||||
{
|
|
||||||
if (!openFileForWrite(cachePath + tocBinFile, tocFile)) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
for (const auto& entry : tocEntries) {
|
|
||||||
writeTocEntry(tocFile, entry);
|
|
||||||
}
|
|
||||||
tocFile.close();
|
|
||||||
}
|
|
||||||
|
|
||||||
// Clear vectors to free memory
|
// Clear vectors to free memory
|
||||||
spineEntries.clear();
|
spineEntries.clear();
|
||||||
spineEntries.shrink_to_fit();
|
spineEntries.shrink_to_fit();
|
||||||
tocEntries.clear();
|
|
||||||
tocEntries.shrink_to_fit();
|
|
||||||
|
|
||||||
Serial.printf("[%lu] [STC] Updated cache with mappings and sizes\n", millis());
|
Serial.printf("[%lu] [STC] Updated cache with mappings and sizes\n", millis());
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Opens (and leaves open all three files for fast access)
|
||||||
bool SpineTocCache::load() {
|
bool SpineTocCache::load() {
|
||||||
// Load metadata
|
// Load metadata
|
||||||
File metaFile;
|
if (!FsHelpers::openFileForRead("STC", cachePath + spineTocMetaBinFile, metaFile)) {
|
||||||
if (!openFileForRead(cachePath + spineTocMetaBinFile, metaFile)) {
|
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -245,10 +245,19 @@ bool SpineTocCache::load() {
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (!FsHelpers::openFileForRead("STC", cachePath + spineBinFile, spineFile)) {
|
||||||
|
metaFile.close();
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!FsHelpers::openFileForRead("STC", cachePath + tocBinFile, tocFile)) {
|
||||||
|
metaFile.close();
|
||||||
|
spineFile.close();
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
serialization::readPod(metaFile, spineCount);
|
serialization::readPod(metaFile, spineCount);
|
||||||
serialization::readPod(metaFile, tocCount);
|
serialization::readPod(metaFile, tocCount);
|
||||||
// TODO: Add LUT to back of meta file
|
|
||||||
metaFile.close();
|
|
||||||
|
|
||||||
loaded = true;
|
loaded = true;
|
||||||
Serial.printf("[%lu] [STC] Loaded cache metadata: %d spine, %d TOC entries\n", millis(), spineCount, tocCount);
|
Serial.printf("[%lu] [STC] Loaded cache metadata: %d spine, %d TOC entries\n", millis(), spineCount, tocCount);
|
||||||
@ -266,18 +275,12 @@ SpineTocCache::SpineEntry SpineTocCache::getSpineEntry(const int index) {
|
|||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!openFileForRead(cachePath + spineBinFile, spineFile)) {
|
// Seek to spine LUT item, read from LUT and get out data
|
||||||
return {};
|
metaFile.seek(SPINE_TOC_META_HEADER_SIZE + sizeof(size_t) * index);
|
||||||
}
|
size_t spineEntryPos;
|
||||||
|
serialization::readPod(metaFile, spineEntryPos);
|
||||||
// Seek to the correct entry - need to read entries sequentially until we reach the index
|
spineFile.seek(spineEntryPos);
|
||||||
// TODO: This could/should be based on a look up table/fixed sizes
|
|
||||||
for (int i = 0; i < index; i++) {
|
|
||||||
readSpineEntry(spineFile); // Skip entries
|
|
||||||
}
|
|
||||||
|
|
||||||
auto entry = readSpineEntry(spineFile);
|
auto entry = readSpineEntry(spineFile);
|
||||||
spineFile.close();
|
|
||||||
return entry;
|
return entry;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -292,18 +295,12 @@ SpineTocCache::TocEntry SpineTocCache::getTocEntry(const int index) {
|
|||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!openFileForRead(cachePath + tocBinFile, tocFile)) {
|
// Seek to TOC LUT item, read from LUT and get out data
|
||||||
return {};
|
metaFile.seek(SPINE_TOC_META_HEADER_SIZE + sizeof(size_t) * spineCount + sizeof(size_t) * index);
|
||||||
}
|
size_t tocEntryPos;
|
||||||
|
serialization::readPod(metaFile, tocEntryPos);
|
||||||
// Seek to the correct entry - need to read entries sequentially until we reach the index
|
tocFile.seek(tocEntryPos);
|
||||||
// TODO: This could/should be based on a look up table/fixed sizes
|
|
||||||
for (int i = 0; i < index; i++) {
|
|
||||||
readTocEntry(tocFile); // Skip entries
|
|
||||||
}
|
|
||||||
|
|
||||||
auto entry = readTocEntry(tocFile);
|
auto entry = readTocEntry(tocFile);
|
||||||
tocFile.close();
|
|
||||||
return entry;
|
return entry;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@ -40,11 +40,12 @@ class SpineTocCache {
|
|||||||
bool buildMode;
|
bool buildMode;
|
||||||
|
|
||||||
// Temp file handles during build
|
// Temp file handles during build
|
||||||
|
File metaFile;
|
||||||
File spineFile;
|
File spineFile;
|
||||||
File tocFile;
|
File tocFile;
|
||||||
|
|
||||||
void writeSpineEntry(File& file, const SpineEntry& entry) const;
|
size_t writeSpineEntry(File& file, const SpineEntry& entry) const;
|
||||||
void writeTocEntry(File& file, const TocEntry& entry) const;
|
size_t writeTocEntry(File& file, const TocEntry& entry) const;
|
||||||
SpineEntry readSpineEntry(File& file) const;
|
SpineEntry readSpineEntry(File& file) const;
|
||||||
TocEntry readTocEntry(File& file) const;
|
TocEntry readTocEntry(File& file) const;
|
||||||
|
|
||||||
@ -60,7 +61,7 @@ class SpineTocCache {
|
|||||||
bool endWrite();
|
bool endWrite();
|
||||||
|
|
||||||
// Post-processing to update mappings and sizes
|
// Post-processing to update mappings and sizes
|
||||||
bool updateMappingsAndSizes(const std::string& epubPath);
|
bool updateMapsAndSizes(const std::string& epubPath);
|
||||||
|
|
||||||
// Reading phase (read mode)
|
// Reading phase (read mode)
|
||||||
bool load();
|
bool load();
|
||||||
|
|||||||
@ -172,7 +172,8 @@ void setup() {
|
|||||||
SPI.begin(EPD_SCLK, SD_SPI_MISO, EPD_MOSI, EPD_CS);
|
SPI.begin(EPD_SCLK, SD_SPI_MISO, EPD_MOSI, EPD_CS);
|
||||||
|
|
||||||
// SD Card Initialization
|
// SD Card Initialization
|
||||||
if (!SD.begin(SD_SPI_CS, SPI, SPI_FQ)) {
|
// We need 6 open files concurrently when parsing a new chapter
|
||||||
|
if (!SD.begin(SD_SPI_CS, SPI, SPI_FQ, "/sd", 6)) {
|
||||||
Serial.printf("[%lu] [ ] SD card initialization failed\n", millis());
|
Serial.printf("[%lu] [ ] SD card initialization failed\n", millis());
|
||||||
exitActivity();
|
exitActivity();
|
||||||
enterNewActivity(new FullScreenMessageActivity(renderer, inputManager, "SD card error", BOLD));
|
enterNewActivity(new FullScreenMessageActivity(renderer, inputManager, "SD card error", BOLD));
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user