mirror of
https://github.com/venthur/blag.git
synced 2025-11-26 05:02:58 +00:00
Compare commits
6 Commits
performanc
...
rebuild_ne
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
60bd86ec22 | ||
|
|
69b729a471 | ||
|
|
f09e841200 | ||
|
|
71c866c9b3 | ||
|
|
c87a300c71 | ||
|
|
a939e6503c |
2
Makefile
2
Makefile
@@ -67,7 +67,7 @@ manpage: $(VENV)
|
|||||||
|
|
||||||
.PHONY: benchmark
|
.PHONY: benchmark
|
||||||
benchmark: $(VENV)
|
benchmark: $(VENV)
|
||||||
$(BIN)/pytest --no-cov -s -rP tests/benchmark.py
|
$(BIN)/pytest --no-cov -capture=no -rA tests/benchmark.py
|
||||||
|
|
||||||
.PHONY: clean
|
.PHONY: clean
|
||||||
clean:
|
clean:
|
||||||
|
|||||||
12
blag/blag.py
12
blag/blag.py
@@ -323,6 +323,18 @@ def process_markdown(
|
|||||||
for src, dst in convertibles:
|
for src, dst in convertibles:
|
||||||
logger.debug(f"Processing {src}")
|
logger.debug(f"Processing {src}")
|
||||||
|
|
||||||
|
# see first if the dst actually needs re-building. for that we compare
|
||||||
|
# the mtimes and assume mtime_dst > mtime_src means that it needs not
|
||||||
|
# to be rebuilt
|
||||||
|
if os.path.exists(f"{output_dir}/{dst}"):
|
||||||
|
mtime_src = os.stat(f"{input_dir}/{src}").st_mtime
|
||||||
|
mtime_dst = os.stat(f"{output_dir}/{dst}").st_mtime
|
||||||
|
if mtime_dst >= mtime_src:
|
||||||
|
logger.debug(
|
||||||
|
"Skipping, as target exists and is newer than source."
|
||||||
|
)
|
||||||
|
continue
|
||||||
|
|
||||||
with open(f"{input_dir}/{src}") as fh:
|
with open(f"{input_dir}/{src}") as fh:
|
||||||
body = fh.read()
|
body = fh.read()
|
||||||
|
|
||||||
|
|||||||
@@ -1,15 +1,24 @@
|
|||||||
|
"""Benchmark the performance of the blag build command."""
|
||||||
|
|
||||||
|
import logging
|
||||||
import os
|
import os
|
||||||
|
from argparse import Namespace
|
||||||
|
|
||||||
|
from pytest import LogCaptureFixture
|
||||||
|
|
||||||
import blag
|
import blag
|
||||||
from blag.blag import build
|
from blag.blag import build
|
||||||
|
|
||||||
|
|
||||||
def test_performance(args) -> None:
|
def test_performance(args: Namespace, caplog: LogCaptureFixture) -> None:
|
||||||
FILES = 1000
|
"""Test performance of the build command."""
|
||||||
print(f"Generating {FILES} files")
|
caplog.set_level(logging.ERROR)
|
||||||
|
|
||||||
|
FILES = 10000
|
||||||
|
print(f"Generating {FILES} markdown files")
|
||||||
# create random markdown files in the content directory
|
# create random markdown files in the content directory
|
||||||
with open(os.path.join(blag.__path__[0], "content", "testpage.md")) as fh:
|
with open(os.path.join(blag.__path__[0], "content", "testpage.md")) as fh:
|
||||||
markdown = fh.read()
|
markdown = fh.read()
|
||||||
for i in range(FILES):
|
for i in range(FILES):
|
||||||
with open(f"content/{i}.md", "w") as f:
|
with open(f"content/{i}.md", "w") as f:
|
||||||
f.write(markdown)
|
f.write(markdown)
|
||||||
@@ -20,12 +29,9 @@ def test_performance(args) -> None:
|
|||||||
t = time()
|
t = time()
|
||||||
build(args)
|
build(args)
|
||||||
t_first = time() - t
|
t_first = time() - t
|
||||||
print(t_first)
|
|
||||||
|
|
||||||
t = time()
|
t = time()
|
||||||
build(args)
|
build(args)
|
||||||
t_second = time() - t
|
t_second = time() - t
|
||||||
print(t_second)
|
|
||||||
print(f"First run: {t_first:.2f}s, second run: {t_second:.2f}s")
|
print(f"First run: {t_first:.2f}s, second run: {t_second:.2f}s")
|
||||||
print(f"Speedup: {t_first/t_second:.2f}")
|
print(f"Speedup: {t_first/t_second:.2f}")
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user