Skip to content

Commit

Permalink
Release 1.0.0
Browse files Browse the repository at this point in the history
  • Loading branch information
darkfrog26 committed Dec 2, 2024
1 parent a33fede commit 046ab2d
Show file tree
Hide file tree
Showing 4 changed files with 14 additions and 15 deletions.
22 changes: 11 additions & 11 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -19,12 +19,12 @@ Computationally focused database using pluggable stores

To add all modules:
```scala
libraryDependencies += "com.outr" %% "lightdb-all" % "0.17.0"
libraryDependencies += "com.outr" %% "lightdb-all" % "1.0.0"
```

For a specific implementation like Lucene:
```scala
libraryDependencies += "com.outr" %% "lightdb-lucene" % "0.17.0"
libraryDependencies += "com.outr" %% "lightdb-lucene" % "1.0.0"
```

## Videos
Expand Down Expand Up @@ -52,7 +52,7 @@ Ensure you have the following:
Add the following dependency to your `build.sbt` file:

```scala
libraryDependencies += "com.outr" %% "lightdb-all" % "0.17.0"
libraryDependencies += "com.outr" %% "lightdb-all" % "1.0.0"
```

---
Expand Down Expand Up @@ -128,7 +128,7 @@ Instantiate and initialize the database:

```scala
val db = new DB
// db: DB = repl.MdocSession$MdocApp$DB@6cd11793
// db: DB = repl.MdocSession$MdocApp$DB@7c34159c
db.init()
// res0: Boolean = true
```
Expand All @@ -145,7 +145,7 @@ val adam = Person(name = "Adam", age = 21)
// city = None,
// nicknames = Set(),
// friends = List(),
// _id = Id(value = "Mg3nibsB1wstqK1xEIiGNeU4Q5iw7Kfc")
// _id = Id(value = "bKaLE07r8AzMoAjBotiP0dDfo9n96I1q")
// )
db.people.transaction { implicit transaction =>
db.people.insert(adam)
Expand All @@ -156,7 +156,7 @@ db.people.transaction { implicit transaction =>
// city = None,
// nicknames = Set(),
// friends = List(),
// _id = Id(value = "Mg3nibsB1wstqK1xEIiGNeU4Q5iw7Kfc")
// _id = Id(value = "bKaLE07r8AzMoAjBotiP0dDfo9n96I1q")
// )
```

Expand All @@ -169,7 +169,7 @@ db.people.transaction { implicit transaction =>
val peopleIn20s = db.people.query.filter(_.age BETWEEN 20 -> 29).toList
println(peopleIn20s)
}
// List(Person(Adam,21,None,Set(),List(),Id(V4HuAlFgFWP0bARChPFtCx5eqMCvtX7l)), Person(Adam,21,None,Set(),List(),Id(xxVmia6PxlkFL1nWWDNFKiO2KccbDrhv)), Person(Adam,21,None,Set(),List(),Id(eaQIHd0ZiDHjxWa9VXzhybMHWtH80C47)), Person(Adam,21,None,Set(),List(),Id(Mg3nibsB1wstqK1xEIiGNeU4Q5iw7Kfc)))
// List(Person(Adam,21,None,Set(),List(),Id(V4HuAlFgFWP0bARChPFtCx5eqMCvtX7l)), Person(Adam,21,None,Set(),List(),Id(xxVmia6PxlkFL1nWWDNFKiO2KccbDrhv)), Person(Adam,21,None,Set(),List(),Id(eaQIHd0ZiDHjxWa9VXzhybMHWtH80C47)), Person(Adam,21,None,Set(),List(),Id(Mg3nibsB1wstqK1xEIiGNeU4Q5iw7Kfc)), Person(Adam,21,None,Set(),List(),Id(bKaLE07r8AzMoAjBotiP0dDfo9n96I1q)))
```

---
Expand Down Expand Up @@ -204,7 +204,7 @@ db.people.transaction { implicit transaction =>
.toList
println(results)
}
// List(MaterializedAggregate({"ageMin": 21, "ageMax": 21, "ageAvg": 21.0, "ageSum": 84},repl.MdocSession$MdocApp$Person$@5e5c66af))
// List(MaterializedAggregate({"ageMin": 21, "ageMax": 21, "ageAvg": 21.0, "ageSum": 105},repl.MdocSession$MdocApp$Person$@7b2728df))
```

### Grouping
Expand All @@ -214,7 +214,7 @@ db.people.transaction { implicit transaction =>
val grouped = db.people.query.grouped(_.age).toList
println(grouped)
}
// List((21,List(Person(Adam,21,None,Set(),List(),Id(V4HuAlFgFWP0bARChPFtCx5eqMCvtX7l)), Person(Adam,21,None,Set(),List(),Id(xxVmia6PxlkFL1nWWDNFKiO2KccbDrhv)), Person(Adam,21,None,Set(),List(),Id(eaQIHd0ZiDHjxWa9VXzhybMHWtH80C47)), Person(Adam,21,None,Set(),List(),Id(Mg3nibsB1wstqK1xEIiGNeU4Q5iw7Kfc)))))
// List((21,List(Person(Adam,21,None,Set(),List(),Id(V4HuAlFgFWP0bARChPFtCx5eqMCvtX7l)), Person(Adam,21,None,Set(),List(),Id(xxVmia6PxlkFL1nWWDNFKiO2KccbDrhv)), Person(Adam,21,None,Set(),List(),Id(eaQIHd0ZiDHjxWa9VXzhybMHWtH80C47)), Person(Adam,21,None,Set(),List(),Id(Mg3nibsB1wstqK1xEIiGNeU4Q5iw7Kfc)), Person(Adam,21,None,Set(),List(),Id(bKaLE07r8AzMoAjBotiP0dDfo9n96I1q)))))
```

---
Expand All @@ -228,14 +228,14 @@ import lightdb.backup._
import java.io.File

DatabaseBackup.archive(db, new File("backup.zip"))
// res5: Int = 5
// res5: Int = 6
```

Restore from a backup:

```scala
DatabaseRestore.archive(db, new File("backup.zip"))
// res6: Int = 5
// res6: Int = 6
```

---
Expand Down
4 changes: 2 additions & 2 deletions build.sbt
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ val developerURL: String = "https://matthicks.com"

name := projectName
ThisBuild / organization := org
ThisBuild / version := "1.0.0-SNAPSHOT"
ThisBuild / version := "1.0.0"
ThisBuild / scalaVersion := scala213
ThisBuild / crossScalaVersions := allScalaVersions
ThisBuild / scalacOptions ++= Seq("-unchecked", "-deprecation")
Expand Down Expand Up @@ -98,7 +98,7 @@ lazy val root = project.in(file("."))
publishLocal := {}
)

lazy val core = crossProject(JVMPlatform) // TODO: Add JSPlatform and NativePlatform
lazy val core = crossProject(JVMPlatform)
.crossType(CrossType.Pure)
.settings(
name := s"$projectName-core",
Expand Down
2 changes: 1 addition & 1 deletion lucene/src/main/scala/lightdb/lucene/LuceneStore.scala
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@ class LuceneStore[Doc <: Document[Doc], Model <: DocumentModel[Doc]](directory:
val latest = Version.LATEST
if (latest != dataVersion) {
// TODO: Support re-indexing
scribe.info(s"Data Version: $dataVersion, Latest Version: $latest")
scribe.warn(s"Data Version: $dataVersion, Latest Version: $latest")
}
}
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,6 @@ case class HikariConnectionManager(config: SQLConfig) extends DataSourceConnecti
config.password.foreach(hc.setPassword)
config.maximumPoolSize.foreach(hc.setMaximumPoolSize)
hc.setAutoCommit(config.autoCommit)
// TODO: Configure
hc.addDataSourceProperty("cachePrepStmts", "true")
hc.addDataSourceProperty("prepStmtCacheSize", "250")
hc.addDataSourceProperty("prepStmtCacheSqlLimit", "2048")
Expand Down

0 comments on commit 046ab2d

Please sign in to comment.