cabal-version: 3.0 name: tokenizer version: 0.1.0.0 synopsis: Check uniqueness and tokenize safely homepage: https://github.com/Lev135/tokenizer description: Provide fast enough uniqueness checking for set of tokens specified on a subset of regular expression. See README for more info. /WARNING this package is not tested enough for the moment. Bugs are very likely here./ license: MIT license-file: LICENSE author: Lev135 maintainer: lev_135@mail.ru copyright: Lev Dvorkin (c) 2022 category: text extra-source-files: CHANGELOG.md extra-doc-files: README.md source-repository head type: git location: https://github.com/Lev135/tokenizer common basics default-language: Haskell2010 default-extensions: DuplicateRecordFields NamedFieldPuns TupleSections RecordWildCards LambdaCase MonadComprehensions GeneralizedNewtypeDeriving StandaloneDeriving DeriveTraversable RankNTypes ScopedTypeVariables PartialTypeSignatures KindSignatures ConstraintKinds DataKinds GADTs TypeFamilies QuantifiedConstraints FlexibleContexts FlexibleInstances TypeOperators TypeApplications ghc-options: -Wall -Wcompat -Wincomplete-record-updates -Wincomplete-uni-patterns -Wredundant-constraints -Wmissing-exported-signatures -Wno-name-shadowing -Wno-unused-do-bind build-depends: base ^>=4.14.3.0 , containers common lib import: basics build-depends: transformers common test import: basics build-depends: megaparsec , hspec library import: lib hs-source-dirs: src exposed-modules: Text.Tokenizer Text.Tokenizer.BlackWhiteSet Text.Tokenizer.Types Text.Tokenizer.Uniqueness Text.Tokenizer.Split test-suite tokenizer-test import: test type: exitcode-stdio-1.0 hs-source-dirs: test main-is: Main.hs build-depends: tokenizer test-suite test-repl import: lib, test type: exitcode-stdio-1.0 hs-source-dirs: src, test main-is: Main.hs other-modules: Text.Tokenizer Text.Tokenizer.BlackWhiteSet Text.Tokenizer.Types Text.Tokenizer.Uniqueness Text.Tokenizer.Split executable examples import: basics hs-source-dirs: examples main-is: Main.hs other-modules: if flag(examples) build-depends: tokenizer , megaparsec else buildable: False flag examples manual: True default: False description: Build examples executable