Compare commits
145 Commits
v2.0.0-beta.3
...
raw
| Author | SHA1 | Date | |
|---|---|---|---|
| b371733c67 | |||
| 64dcce07ea | |||
| 28f1efc7d3 | |||
| 7d69e4a728 | |||
| e46d245c09 | |||
| 7baa23f493 | |||
| 2d8433b69e | |||
| 67bc5422f3 | |||
| fb6d1d6c2b | |||
| d017a6dc89 | |||
| d6d3196163 | |||
| 41718a6db3 | |||
| 216628222f | |||
| 322e0b15d2 | |||
| 85bfc0ed51 | |||
| 295a720dfb | |||
| 0a422e3dbd | |||
| 627dade0c7 | |||
| b2e0231cc9 | |||
| ba95863cd3 | |||
| db679df765 | |||
| c5ca2c682b | |||
| ed80712cb4 | |||
| b24772942d | |||
| 9501a05ed7 | |||
| 171a592663 | |||
| 5aaf5ef13b | |||
| adacebd8c7 | |||
| 8bbb673431 | |||
| 2377ac4bc0 | |||
| f5cc8c49eb | |||
| 89d7b412d8 | |||
| 88a8aecdd4 | |||
| 9804fc57e0 | |||
| 068279f13b | |||
| b9edbeb611 | |||
| a97c9317d4 | |||
| 3229a0abfb | |||
| 3f5d8a6b06 | |||
| 146f70ea8a | |||
| e83cf535f5 | |||
| c3ba3ef97a | |||
| 7ee3c8ff25 | |||
| 1e85aa6d78 | |||
| 46fa3225e2 | |||
| 4d51831dab | |||
| 5a1a96cb2d | |||
| ea9040ae83 | |||
| 2373685f1e | |||
| f1391952d4 | |||
| 4a73a200ed | |||
| 4807229e94 | |||
| d8ddc00c61 | |||
| 82f8dad811 | |||
| 75db1016e8 | |||
| de6d715bd2 | |||
| 3ab2fc2b87 | |||
| 1b1dd3d6d5 | |||
| 128b7a8bfb | |||
| 892df5c28e | |||
| d58eb50ebf | |||
| 535fc65c5f | |||
| f158d7d278 | |||
| 5fd6e9cce0 | |||
| 8ce5c3d78f | |||
| 177b4a5e53 | |||
| 5cbdea6192 | |||
| 696dd25c17 | |||
| facb2b13e8 | |||
| 8bbb519477 | |||
| b37e11d74d | |||
| 6cd86876b8 | |||
| f53bc740c1 | |||
| 9bf9be681e | |||
| c862c344b3 | |||
| 0d20a84523 | |||
| 3990899d7e | |||
| 4c7a337083 | |||
| bbaae540ce | |||
| ede6445608 | |||
| b226db6a29 | |||
| d8997efb5a | |||
| 79e78b234c | |||
| 1b5a25c0ef | |||
| 8eae15b2ee | |||
| 2b3de620e8 | |||
| 8645d6376b | |||
| 64fe47161f | |||
| 4dff8eaa4d | |||
| 2dbd29a565 | |||
| f27a07d31a | |||
| 644515958c | |||
| 8683be35f6 | |||
| dc1740d473 | |||
| 11f789ef11 | |||
| 74d21b367f | |||
| 6617e7e73d | |||
| 3dbca20bc9 | |||
| 85c0658984 | |||
| 772d169b52 | |||
| b4ec220f7e | |||
| 3694ae88f6 | |||
| 19751e8a51 | |||
| 925f214125 | |||
| 39f893ad99 | |||
| c871a61015 | |||
| d0d001625c | |||
| 64941b99e2 | |||
| ed02a1f192 | |||
| 4d7c9ddac7 | |||
| feb1830dcc | |||
| 1c33d6ce20 | |||
| 3000471a12 | |||
| 1f33a6a476 | |||
| 2700aad5d2 | |||
| 7ccaa2744e | |||
| df4bb061f8 | |||
| 9e81ce1c33 | |||
| a23850f29b | |||
| 76f53c857b | |||
| 85f5d567e4 | |||
| bd5cba0b0b | |||
| cd54472d03 | |||
| cc0d1a90ff | |||
| 4984dcb5e9 | |||
| 86632bc190 | |||
| d25eec183f | |||
| e96746311c | |||
| 62acca2b68 | |||
| 476492a85c | |||
| ee9b902222 | |||
| fa56f48daf | |||
| f34c9c332f | |||
| a0d685d482 | |||
| 4a5ae9e81e | |||
| 7e2fa1bc80 | |||
| 40cfb6f458 | |||
| 1230ca485e | |||
| 69ab7e10d1 | |||
| fa07960695 | |||
| 8be357dfa1 | |||
| a93b34d984 | |||
| 9c24fbeaad | |||
| f6b38c33b7 | |||
| 773f10110c |
@@ -1,3 +1,4 @@
|
|||||||
* text=auto
|
* text=auto
|
||||||
|
|
||||||
benchmark/benchmark.toml text eol=lf
|
benchmark/benchmark.toml text eol=lf
|
||||||
|
testdata/** text eol=lf
|
||||||
|
|||||||
+15
-4
@@ -1,6 +1,17 @@
|
|||||||
version: 2
|
version: 2
|
||||||
updates:
|
updates:
|
||||||
- package-ecosystem: "gomod"
|
- package-ecosystem: gomod
|
||||||
directory: "/" # Location of package manifests
|
directory: /
|
||||||
schedule:
|
schedule:
|
||||||
interval: "daily"
|
interval: daily
|
||||||
|
open-pull-requests-limit: 10
|
||||||
|
- package-ecosystem: github-actions
|
||||||
|
directory: /
|
||||||
|
schedule:
|
||||||
|
interval: daily
|
||||||
|
open-pull-requests-limit: 10
|
||||||
|
- package-ecosystem: docker
|
||||||
|
directory: /
|
||||||
|
schedule:
|
||||||
|
interval: daily
|
||||||
|
open-pull-requests-limit: 10
|
||||||
|
|||||||
@@ -0,0 +1,21 @@
|
|||||||
|
changelog:
|
||||||
|
exclude:
|
||||||
|
labels:
|
||||||
|
- build
|
||||||
|
- testing
|
||||||
|
categories:
|
||||||
|
- title: What's new
|
||||||
|
labels:
|
||||||
|
- feature
|
||||||
|
- title: Performance
|
||||||
|
labels:
|
||||||
|
- performance
|
||||||
|
- title: Fixed bugs
|
||||||
|
labels:
|
||||||
|
- bug
|
||||||
|
- title: Documentation
|
||||||
|
labels:
|
||||||
|
- doc
|
||||||
|
- title: Other changes
|
||||||
|
labels:
|
||||||
|
- "*"
|
||||||
@@ -35,11 +35,11 @@ jobs:
|
|||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout repository
|
- name: Checkout repository
|
||||||
uses: actions/checkout@v2
|
uses: actions/checkout@v3
|
||||||
|
|
||||||
# Initializes the CodeQL tools for scanning.
|
# Initializes the CodeQL tools for scanning.
|
||||||
- name: Initialize CodeQL
|
- name: Initialize CodeQL
|
||||||
uses: github/codeql-action/init@v1
|
uses: github/codeql-action/init@v2
|
||||||
with:
|
with:
|
||||||
languages: ${{ matrix.language }}
|
languages: ${{ matrix.language }}
|
||||||
# If you wish to specify custom queries, you can do so here or in a config file.
|
# If you wish to specify custom queries, you can do so here or in a config file.
|
||||||
@@ -50,7 +50,7 @@ jobs:
|
|||||||
# Autobuild attempts to build any compiled languages (C/C++, C#, or Java).
|
# Autobuild attempts to build any compiled languages (C/C++, C#, or Java).
|
||||||
# If this step fails, then you should remove it and run the build manually (see below)
|
# If this step fails, then you should remove it and run the build manually (see below)
|
||||||
- name: Autobuild
|
- name: Autobuild
|
||||||
uses: github/codeql-action/autobuild@v1
|
uses: github/codeql-action/autobuild@v2
|
||||||
|
|
||||||
# ℹ️ Command-line programs to run using the OS shell.
|
# ℹ️ Command-line programs to run using the OS shell.
|
||||||
# 📚 https://git.io/JvXDl
|
# 📚 https://git.io/JvXDl
|
||||||
@@ -64,4 +64,4 @@ jobs:
|
|||||||
# make release
|
# make release
|
||||||
|
|
||||||
- name: Perform CodeQL Analysis
|
- name: Perform CodeQL Analysis
|
||||||
uses: github/codeql-action/analyze@v1
|
uses: github/codeql-action/analyze@v2
|
||||||
|
|||||||
@@ -9,12 +9,12 @@ jobs:
|
|||||||
runs-on: "ubuntu-latest"
|
runs-on: "ubuntu-latest"
|
||||||
name: report
|
name: report
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@master
|
- uses: actions/checkout@v3
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
- name: Setup go
|
- name: Setup go
|
||||||
uses: actions/setup-go@master
|
uses: actions/setup-go@master
|
||||||
with:
|
with:
|
||||||
go-version: 1.16
|
go-version: 1.19
|
||||||
- name: Run tests with coverage
|
- name: Run tests with coverage
|
||||||
run: ./ci.sh coverage -d "${GITHUB_BASE_REF-HEAD}"
|
run: ./ci.sh coverage -d "${GITHUB_BASE_REF-HEAD}"
|
||||||
|
|||||||
@@ -0,0 +1,39 @@
|
|||||||
|
name: release
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
tags:
|
||||||
|
- "v2.*"
|
||||||
|
workflow_call:
|
||||||
|
inputs:
|
||||||
|
args:
|
||||||
|
description: "Extra arguments to pass goreleaser"
|
||||||
|
default: ""
|
||||||
|
required: false
|
||||||
|
type: string
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
release:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
steps:
|
||||||
|
- name: Checkout
|
||||||
|
uses: actions/checkout@v3
|
||||||
|
with:
|
||||||
|
fetch-depth: 0
|
||||||
|
- name: Set up Go
|
||||||
|
uses: actions/setup-go@v2
|
||||||
|
with:
|
||||||
|
go-version: 1.19
|
||||||
|
- name: Login to GitHub Container Registry
|
||||||
|
uses: docker/login-action@v2
|
||||||
|
with:
|
||||||
|
registry: ghcr.io
|
||||||
|
username: ${{ github.actor }}
|
||||||
|
password: ${{ secrets.GITHUB_TOKEN }}
|
||||||
|
- name: Run GoReleaser
|
||||||
|
uses: goreleaser/goreleaser-action@v3
|
||||||
|
with:
|
||||||
|
distribution: goreleaser
|
||||||
|
version: latest
|
||||||
|
args: release ${{ inputs.args }} --rm-dist
|
||||||
|
env:
|
||||||
|
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
||||||
@@ -12,14 +12,21 @@ jobs:
|
|||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
os: [ 'ubuntu-latest', 'windows-latest', 'macos-latest']
|
os: [ 'ubuntu-latest', 'windows-latest', 'macos-latest']
|
||||||
go: [ '1.15', '1.16' ]
|
go: [ '1.18', '1.19' ]
|
||||||
runs-on: ${{ matrix.os }}
|
runs-on: ${{ matrix.os }}
|
||||||
name: ${{ matrix.go }}/${{ matrix.os }}
|
name: ${{ matrix.go }}/${{ matrix.os }}
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@master
|
- uses: actions/checkout@v3
|
||||||
|
with:
|
||||||
|
fetch-depth: 0
|
||||||
- name: Setup go ${{ matrix.go }}
|
- name: Setup go ${{ matrix.go }}
|
||||||
uses: actions/setup-go@master
|
uses: actions/setup-go@master
|
||||||
with:
|
with:
|
||||||
go-version: ${{ matrix.go }}
|
go-version: ${{ matrix.go }}
|
||||||
- name: Run unit tests
|
- name: Run unit tests
|
||||||
run: go test -race ./...
|
run: go test -race ./...
|
||||||
|
release-check:
|
||||||
|
if: ${{ github.ref != 'refs/heads/v2' }}
|
||||||
|
uses: pelletier/go-toml/.github/workflows/release.yml@v2
|
||||||
|
with:
|
||||||
|
args: --snapshot
|
||||||
|
|||||||
@@ -3,3 +3,4 @@ fuzz/
|
|||||||
cmd/tomll/tomll
|
cmd/tomll/tomll
|
||||||
cmd/tomljson/tomljson
|
cmd/tomljson/tomljson
|
||||||
cmd/tomltestgen/tomltestgen
|
cmd/tomltestgen/tomltestgen
|
||||||
|
dist
|
||||||
@@ -0,0 +1,123 @@
|
|||||||
|
before:
|
||||||
|
hooks:
|
||||||
|
- go mod tidy
|
||||||
|
- go fmt ./...
|
||||||
|
- go test ./...
|
||||||
|
builds:
|
||||||
|
- id: tomll
|
||||||
|
main: ./cmd/tomll
|
||||||
|
binary: tomll
|
||||||
|
env:
|
||||||
|
- CGO_ENABLED=0
|
||||||
|
flags:
|
||||||
|
- -trimpath
|
||||||
|
ldflags:
|
||||||
|
- -X main.version={{.Version}} -X main.commit={{.Commit}} -X main.date={{.CommitDate}}
|
||||||
|
mod_timestamp: '{{ .CommitTimestamp }}'
|
||||||
|
targets:
|
||||||
|
- linux_amd64
|
||||||
|
- linux_arm64
|
||||||
|
- linux_arm
|
||||||
|
- windows_amd64
|
||||||
|
- windows_arm64
|
||||||
|
- windows_arm
|
||||||
|
- darwin_amd64
|
||||||
|
- darwin_arm64
|
||||||
|
- id: tomljson
|
||||||
|
main: ./cmd/tomljson
|
||||||
|
binary: tomljson
|
||||||
|
env:
|
||||||
|
- CGO_ENABLED=0
|
||||||
|
flags:
|
||||||
|
- -trimpath
|
||||||
|
ldflags:
|
||||||
|
- -X main.version={{.Version}} -X main.commit={{.Commit}} -X main.date={{.CommitDate}}
|
||||||
|
mod_timestamp: '{{ .CommitTimestamp }}'
|
||||||
|
targets:
|
||||||
|
- linux_amd64
|
||||||
|
- linux_arm64
|
||||||
|
- linux_arm
|
||||||
|
- windows_amd64
|
||||||
|
- windows_arm64
|
||||||
|
- windows_arm
|
||||||
|
- darwin_amd64
|
||||||
|
- darwin_arm64
|
||||||
|
- id: jsontoml
|
||||||
|
main: ./cmd/jsontoml
|
||||||
|
binary: jsontoml
|
||||||
|
env:
|
||||||
|
- CGO_ENABLED=0
|
||||||
|
flags:
|
||||||
|
- -trimpath
|
||||||
|
ldflags:
|
||||||
|
- -X main.version={{.Version}} -X main.commit={{.Commit}} -X main.date={{.CommitDate}}
|
||||||
|
mod_timestamp: '{{ .CommitTimestamp }}'
|
||||||
|
targets:
|
||||||
|
- linux_amd64
|
||||||
|
- linux_arm64
|
||||||
|
- linux_arm
|
||||||
|
- windows_amd64
|
||||||
|
- windows_arm64
|
||||||
|
- windows_arm
|
||||||
|
- darwin_amd64
|
||||||
|
- darwin_arm64
|
||||||
|
universal_binaries:
|
||||||
|
- id: tomll
|
||||||
|
replace: true
|
||||||
|
name_template: tomll
|
||||||
|
- id: tomljson
|
||||||
|
replace: true
|
||||||
|
name_template: tomljson
|
||||||
|
- id: jsontoml
|
||||||
|
replace: true
|
||||||
|
name_template: jsontoml
|
||||||
|
archives:
|
||||||
|
- id: jsontoml
|
||||||
|
format: tar.xz
|
||||||
|
builds:
|
||||||
|
- jsontoml
|
||||||
|
files:
|
||||||
|
- none*
|
||||||
|
name_template: "{{ .Binary }}_{{.Version}}_{{ .Os }}_{{ .Arch }}"
|
||||||
|
- id: tomljson
|
||||||
|
format: tar.xz
|
||||||
|
builds:
|
||||||
|
- tomljson
|
||||||
|
files:
|
||||||
|
- none*
|
||||||
|
name_template: "{{ .Binary }}_{{.Version}}_{{ .Os }}_{{ .Arch }}"
|
||||||
|
- id: tomll
|
||||||
|
format: tar.xz
|
||||||
|
builds:
|
||||||
|
- tomll
|
||||||
|
files:
|
||||||
|
- none*
|
||||||
|
name_template: "{{ .Binary }}_{{.Version}}_{{ .Os }}_{{ .Arch }}"
|
||||||
|
dockers:
|
||||||
|
- id: tools
|
||||||
|
goos: linux
|
||||||
|
goarch: amd64
|
||||||
|
ids:
|
||||||
|
- jsontoml
|
||||||
|
- tomljson
|
||||||
|
- tomll
|
||||||
|
image_templates:
|
||||||
|
- "ghcr.io/pelletier/go-toml:latest"
|
||||||
|
- "ghcr.io/pelletier/go-toml:{{ .Tag }}"
|
||||||
|
- "ghcr.io/pelletier/go-toml:v{{ .Major }}"
|
||||||
|
skip_push: false
|
||||||
|
checksum:
|
||||||
|
name_template: 'sha256sums.txt'
|
||||||
|
snapshot:
|
||||||
|
name_template: "{{ incpatch .Version }}-next"
|
||||||
|
release:
|
||||||
|
github:
|
||||||
|
owner: pelletier
|
||||||
|
name: go-toml
|
||||||
|
draft: true
|
||||||
|
prerelease: auto
|
||||||
|
mode: replace
|
||||||
|
changelog:
|
||||||
|
use: github-native
|
||||||
|
announce:
|
||||||
|
skip: true
|
||||||
+23
-9
@@ -155,6 +155,8 @@ Checklist:
|
|||||||
- Does not introduce backward-incompatible changes (unless discussed).
|
- Does not introduce backward-incompatible changes (unless discussed).
|
||||||
- Has relevant doc changes.
|
- Has relevant doc changes.
|
||||||
- Benchstat does not show performance regression.
|
- Benchstat does not show performance regression.
|
||||||
|
- Pull request is [labeled appropriately][pr-labels].
|
||||||
|
- Title will be understandable in the changelog.
|
||||||
|
|
||||||
1. Merge using "squash and merge".
|
1. Merge using "squash and merge".
|
||||||
2. Make sure to edit the commit message to keep all the useful information
|
2. Make sure to edit the commit message to keep all the useful information
|
||||||
@@ -163,13 +165,25 @@ Checklist:
|
|||||||
|
|
||||||
### New release
|
### New release
|
||||||
|
|
||||||
1. Go to [releases][releases]. Click on "X commits to master since this
|
1. Decide on the next version number. Use semver.
|
||||||
release".
|
2. Generate release notes using [`gh`][gh]. Example:
|
||||||
2. Make note of all the changes. Look for backward incompatible changes,
|
```
|
||||||
new features, and bug fixes.
|
$ gh api -X POST \
|
||||||
3. Pick the new version using the above and semver.
|
-F tag_name='v2.0.0-beta.5' \
|
||||||
4. Create a [new release][new-release].
|
-F target_commitish='v2' \
|
||||||
5. Follow the same format as [1.1.0][release-110].
|
-F previous_tag_name='v2.0.0-beta.4' \
|
||||||
|
--jq '.body' \
|
||||||
|
repos/pelletier/go-toml/releases/generate-notes
|
||||||
|
```
|
||||||
|
3. Look for "Other changes". That would indicate a pull request not labeled
|
||||||
|
properly. Tweak labels and pull request titles until changelog looks good for
|
||||||
|
users.
|
||||||
|
4. [Draft new release][new-release].
|
||||||
|
5. Fill tag and target with the same value used to generate the changelog.
|
||||||
|
6. Set title to the new tag value.
|
||||||
|
7. Paste the generated changelog.
|
||||||
|
8. Check "create discussion", in the "Releases" category.
|
||||||
|
9. Check pre-release if new version is an alpha or beta.
|
||||||
|
|
||||||
[issues-tracker]: https://github.com/pelletier/go-toml/issues
|
[issues-tracker]: https://github.com/pelletier/go-toml/issues
|
||||||
[bug-report]: https://github.com/pelletier/go-toml/issues/new?template=bug_report.md
|
[bug-report]: https://github.com/pelletier/go-toml/issues/new?template=bug_report.md
|
||||||
@@ -177,6 +191,6 @@ Checklist:
|
|||||||
[readme]: ./README.md
|
[readme]: ./README.md
|
||||||
[fork]: https://help.github.com/articles/fork-a-repo
|
[fork]: https://help.github.com/articles/fork-a-repo
|
||||||
[pull-request]: https://help.github.com/en/articles/creating-a-pull-request
|
[pull-request]: https://help.github.com/en/articles/creating-a-pull-request
|
||||||
[releases]: https://github.com/pelletier/go-toml/releases
|
|
||||||
[new-release]: https://github.com/pelletier/go-toml/releases/new
|
[new-release]: https://github.com/pelletier/go-toml/releases/new
|
||||||
[release-110]: https://github.com/pelletier/go-toml/releases/tag/v1.1.0
|
[gh]: https://github.com/cli/cli
|
||||||
|
[pr-labels]: https://github.com/pelletier/go-toml/blob/v2/.github/release.yml
|
||||||
|
|||||||
@@ -0,0 +1,5 @@
|
|||||||
|
FROM scratch
|
||||||
|
ENV PATH "$PATH:/bin"
|
||||||
|
COPY tomll /bin/tomll
|
||||||
|
COPY tomljson /bin/tomljson
|
||||||
|
COPY jsontoml /bin/jsontoml
|
||||||
@@ -1,6 +1,6 @@
|
|||||||
The MIT License (MIT)
|
The MIT License (MIT)
|
||||||
|
|
||||||
Copyright (c) 2013 - 2021 Thomas Pelletier, Eric Anderton
|
Copyright (c) 2013 - 2022 Thomas Pelletier, Eric Anderton
|
||||||
|
|
||||||
Permission is hereby granted, free of charge, to any person obtaining a copy
|
Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||||
of this software and associated documentation files (the "Software"), to deal
|
of this software and associated documentation files (the "Software"), to deal
|
||||||
|
|||||||
@@ -4,24 +4,14 @@ Go library for the [TOML](https://toml.io/en/) format.
|
|||||||
|
|
||||||
This library supports [TOML v1.0.0](https://toml.io/en/v1.0.0).
|
This library supports [TOML v1.0.0](https://toml.io/en/v1.0.0).
|
||||||
|
|
||||||
## Development status
|
|
||||||
|
|
||||||
This is the upcoming major version of go-toml. It is currently in active
|
|
||||||
development. As of release v2.0.0-beta.1, the library has reached feature parity
|
|
||||||
with v1, and fixes a lot known bugs and performance issues along the way.
|
|
||||||
|
|
||||||
If you do not need the advanced document editing features of v1, you are
|
|
||||||
encouraged to try out this version.
|
|
||||||
|
|
||||||
[👉 Roadmap for v2](https://github.com/pelletier/go-toml/discussions/506)
|
|
||||||
|
|
||||||
[🐞 Bug Reports](https://github.com/pelletier/go-toml/issues)
|
[🐞 Bug Reports](https://github.com/pelletier/go-toml/issues)
|
||||||
|
|
||||||
[💬 Anything else](https://github.com/pelletier/go-toml/discussions)
|
[💬 Anything else](https://github.com/pelletier/go-toml/discussions)
|
||||||
|
|
||||||
## Documentation
|
## Documentation
|
||||||
|
|
||||||
Full API, examples, and implementation notes are available in the Go documentation.
|
Full API, examples, and implementation notes are available in the Go
|
||||||
|
documentation.
|
||||||
|
|
||||||
[](https://pkg.go.dev/github.com/pelletier/go-toml/v2)
|
[](https://pkg.go.dev/github.com/pelletier/go-toml/v2)
|
||||||
|
|
||||||
@@ -31,6 +21,8 @@ Full API, examples, and implementation notes are available in the Go documentati
|
|||||||
import "github.com/pelletier/go-toml/v2"
|
import "github.com/pelletier/go-toml/v2"
|
||||||
```
|
```
|
||||||
|
|
||||||
|
See [Modules](#Modules).
|
||||||
|
|
||||||
## Features
|
## Features
|
||||||
|
|
||||||
### Stdlib behavior
|
### Stdlib behavior
|
||||||
@@ -46,15 +38,16 @@ operations should not be shockingly slow. See [benchmarks](#benchmarks).
|
|||||||
### Strict mode
|
### Strict mode
|
||||||
|
|
||||||
`Decoder` can be set to "strict mode", which makes it error when some parts of
|
`Decoder` can be set to "strict mode", which makes it error when some parts of
|
||||||
the TOML document was not prevent in the target structure. This is a great way
|
the TOML document was not present in the target structure. This is a great way
|
||||||
to check for typos. [See example in the documentation][strict].
|
to check for typos. [See example in the documentation][strict].
|
||||||
|
|
||||||
[strict]: https://pkg.go.dev/github.com/pelletier/go-toml/v2#example-Decoder.SetStrict
|
[strict]: https://pkg.go.dev/github.com/pelletier/go-toml/v2#example-Decoder.DisallowUnknownFields
|
||||||
|
|
||||||
### Contextualized errors
|
### Contextualized errors
|
||||||
|
|
||||||
When decoding errors occur, go-toml returns [`DecodeError`][decode-err]), which
|
When most decoding errors occur, go-toml returns [`DecodeError`][decode-err]),
|
||||||
contains a human readable contextualized version of the error. For example:
|
which contains a human readable contextualized version of the error. For
|
||||||
|
example:
|
||||||
|
|
||||||
```
|
```
|
||||||
2| key1 = "value1"
|
2| key1 = "value1"
|
||||||
@@ -156,12 +149,12 @@ Execution time speedup compared to other Go TOML libraries:
|
|||||||
<tr><th>Benchmark</th><th>go-toml v1</th><th>BurntSushi/toml</th></tr>
|
<tr><th>Benchmark</th><th>go-toml v1</th><th>BurntSushi/toml</th></tr>
|
||||||
</thead>
|
</thead>
|
||||||
<tbody>
|
<tbody>
|
||||||
<tr><td>Marshal/HugoFrontMatter</td><td>2.0x</td><td>2.0x</td></tr>
|
<tr><td>Marshal/HugoFrontMatter-2</td><td>1.9x</td><td>1.9x</td></tr>
|
||||||
<tr><td>Marshal/ReferenceFile/map</td><td>1.8x</td><td>2.0x</td></tr>
|
<tr><td>Marshal/ReferenceFile/map-2</td><td>1.7x</td><td>1.8x</td></tr>
|
||||||
<tr><td>Marshal/ReferenceFile/struct</td><td>2.7x</td><td>2.7x</td></tr>
|
<tr><td>Marshal/ReferenceFile/struct-2</td><td>2.2x</td><td>2.5x</td></tr>
|
||||||
<tr><td>Unmarshal/HugoFrontMatter</td><td>3.0x</td><td>2.6x</td></tr>
|
<tr><td>Unmarshal/HugoFrontMatter-2</td><td>2.9x</td><td>2.9x</td></tr>
|
||||||
<tr><td>Unmarshal/ReferenceFile/map</td><td>3.0x</td><td>3.1x</td></tr>
|
<tr><td>Unmarshal/ReferenceFile/map-2</td><td>2.6x</td><td>2.9x</td></tr>
|
||||||
<tr><td>Unmarshal/ReferenceFile/struct</td><td>5.9x</td><td>6.6x</td></tr>
|
<tr><td>Unmarshal/ReferenceFile/struct-2</td><td>4.4x</td><td>5.3x</td></tr>
|
||||||
</tbody>
|
</tbody>
|
||||||
</table>
|
</table>
|
||||||
<details><summary>See more</summary>
|
<details><summary>See more</summary>
|
||||||
@@ -174,21 +167,74 @@ provided for completeness.</p>
|
|||||||
<tr><th>Benchmark</th><th>go-toml v1</th><th>BurntSushi/toml</th></tr>
|
<tr><th>Benchmark</th><th>go-toml v1</th><th>BurntSushi/toml</th></tr>
|
||||||
</thead>
|
</thead>
|
||||||
<tbody>
|
<tbody>
|
||||||
<tr><td>Marshal/SimpleDocument/map</td><td>1.7x</td><td>2.1x</td></tr>
|
<tr><td>Marshal/SimpleDocument/map-2</td><td>1.8x</td><td>2.9x</td></tr>
|
||||||
<tr><td>Marshal/SimpleDocument/struct</td><td>2.6x</td><td>2.9x</td></tr>
|
<tr><td>Marshal/SimpleDocument/struct-2</td><td>2.7x</td><td>4.2x</td></tr>
|
||||||
<tr><td>Unmarshal/SimpleDocument/map</td><td>4.1x</td><td>2.9x</td></tr>
|
<tr><td>Unmarshal/SimpleDocument/map-2</td><td>4.5x</td><td>3.1x</td></tr>
|
||||||
<tr><td>Unmarshal/SimpleDocument/struct</td><td>6.3x</td><td>4.1x</td></tr>
|
<tr><td>Unmarshal/SimpleDocument/struct-2</td><td>6.2x</td><td>3.9x</td></tr>
|
||||||
<tr><td>UnmarshalDataset/example</td><td>3.5x</td><td>2.4x</td></tr>
|
<tr><td>UnmarshalDataset/example-2</td><td>3.1x</td><td>3.5x</td></tr>
|
||||||
<tr><td>UnmarshalDataset/code</td><td>2.2x</td><td>2.8x</td></tr>
|
<tr><td>UnmarshalDataset/code-2</td><td>2.3x</td><td>3.1x</td></tr>
|
||||||
<tr><td>UnmarshalDataset/twitter</td><td>2.8x</td><td>2.1x</td></tr>
|
<tr><td>UnmarshalDataset/twitter-2</td><td>2.5x</td><td>2.6x</td></tr>
|
||||||
<tr><td>UnmarshalDataset/citm_catalog</td><td>2.3x</td><td>1.5x</td></tr>
|
<tr><td>UnmarshalDataset/citm_catalog-2</td><td>2.1x</td><td>2.2x</td></tr>
|
||||||
<tr><td>UnmarshalDataset/config</td><td>4.2x</td><td>3.2x</td></tr>
|
<tr><td>UnmarshalDataset/canada-2</td><td>1.6x</td><td>1.3x</td></tr>
|
||||||
<tr><td>[Geo mean]</td><td>3.0x</td><td>2.7x</td></tr>
|
<tr><td>UnmarshalDataset/config-2</td><td>4.3x</td><td>3.2x</td></tr>
|
||||||
|
<tr><td>[Geo mean]</td><td>2.7x</td><td>2.8x</td></tr>
|
||||||
</tbody>
|
</tbody>
|
||||||
</table>
|
</table>
|
||||||
<p>This table can be generated with <code>./ci.sh benchmark -a -html</code>.</p>
|
<p>This table can be generated with <code>./ci.sh benchmark -a -html</code>.</p>
|
||||||
</details>
|
</details>
|
||||||
|
|
||||||
|
## Modules
|
||||||
|
|
||||||
|
go-toml uses Go's standard modules system.
|
||||||
|
|
||||||
|
Installation instructions:
|
||||||
|
|
||||||
|
- Go ≥ 1.16: Nothing to do. Use the import in your code. The `go` command deals
|
||||||
|
with it automatically.
|
||||||
|
- Go ≥ 1.13: `GO111MODULE=on go get github.com/pelletier/go-toml/v2`.
|
||||||
|
|
||||||
|
In case of trouble: [Go Modules FAQ][mod-faq].
|
||||||
|
|
||||||
|
[mod-faq]: https://github.com/golang/go/wiki/Modules#why-does-installing-a-tool-via-go-get-fail-with-error-cannot-find-main-module
|
||||||
|
|
||||||
|
## Tools
|
||||||
|
|
||||||
|
Go-toml provides three handy command line tools:
|
||||||
|
|
||||||
|
* `tomljson`: Reads a TOML file and outputs its JSON representation.
|
||||||
|
|
||||||
|
```
|
||||||
|
$ go install github.com/pelletier/go-toml/v2/cmd/tomljson@latest
|
||||||
|
$ tomljson --help
|
||||||
|
```
|
||||||
|
|
||||||
|
* `jsontoml`: Reads a JSON file and outputs a TOML representation.
|
||||||
|
|
||||||
|
```
|
||||||
|
$ go install github.com/pelletier/go-toml/v2/cmd/jsontoml@latest
|
||||||
|
$ jsontoml --help
|
||||||
|
```
|
||||||
|
|
||||||
|
* `tomll`: Lints and reformats a TOML file.
|
||||||
|
|
||||||
|
```
|
||||||
|
$ go install github.com/pelletier/go-toml/v2/cmd/tomll@latest
|
||||||
|
$ tomll --help
|
||||||
|
```
|
||||||
|
|
||||||
|
### Docker image
|
||||||
|
|
||||||
|
Those tools are also available as a [Docker image][docker]. For example, to use
|
||||||
|
`tomljson`:
|
||||||
|
|
||||||
|
```
|
||||||
|
docker run -i ghcr.io/pelletier/go-toml:v2 tomljson < example.toml
|
||||||
|
```
|
||||||
|
|
||||||
|
Multiple versions are availble on [ghcr.io][docker].
|
||||||
|
|
||||||
|
[docker]: https://github.com/pelletier/go-toml/pkgs/container/go-toml
|
||||||
|
|
||||||
## Migrating from v1
|
## Migrating from v1
|
||||||
|
|
||||||
This section describes the differences between v1 and v2, with some pointers on
|
This section describes the differences between v1 and v2, with some pointers on
|
||||||
@@ -307,6 +353,29 @@ The recommended replacement is pre-filling the struct before unmarshaling.
|
|||||||
|
|
||||||
[go-defaults]: https://github.com/mcuadros/go-defaults
|
[go-defaults]: https://github.com/mcuadros/go-defaults
|
||||||
|
|
||||||
|
#### `toml.Tree` replacement
|
||||||
|
|
||||||
|
This structure was the initial attempt at providing a document model for
|
||||||
|
go-toml. It allows manipulating the structure of any document, encoding and
|
||||||
|
decoding from their TOML representation. While a more robust feature was
|
||||||
|
initially planned in go-toml v2, this has been ultimately [removed from
|
||||||
|
scope][nodoc] of this library, with no plan to add it back at the moment. The
|
||||||
|
closest equivalent at the moment would be to unmarshal into an `interface{}` and
|
||||||
|
use type assertions and/or reflection to manipulate the arbitrary
|
||||||
|
structure. However this would fall short of providing all of the TOML features
|
||||||
|
such as adding comments and be specific about whitespace.
|
||||||
|
|
||||||
|
|
||||||
|
#### `toml.Position` are not retrievable anymore
|
||||||
|
|
||||||
|
The API for retrieving the position (line, column) of a specific TOML element do
|
||||||
|
not exist anymore. This was done to minimize the amount of concepts introduced
|
||||||
|
by the library (query path), and avoid the performance hit related to storing
|
||||||
|
positions in the absence of a document model, for a feature that seemed to have
|
||||||
|
little use. Errors however have gained more detailed position
|
||||||
|
information. Position retrieval seems better fitted for a document model, which
|
||||||
|
has been [removed from the scope][nodoc] of go-toml v2 at the moment.
|
||||||
|
|
||||||
### Encoding / Marshal
|
### Encoding / Marshal
|
||||||
|
|
||||||
#### Default struct fields order
|
#### Default struct fields order
|
||||||
@@ -342,7 +411,8 @@ fmt.Println("v2:\n" + string(b))
|
|||||||
```
|
```
|
||||||
|
|
||||||
There is no way to make v2 encoder behave like v1. A workaround could be to
|
There is no way to make v2 encoder behave like v1. A workaround could be to
|
||||||
manually sort the fields alphabetically in the struct definition.
|
manually sort the fields alphabetically in the struct definition, or generate
|
||||||
|
struct types using `reflect.StructOf`.
|
||||||
|
|
||||||
#### No indentation by default
|
#### No indentation by default
|
||||||
|
|
||||||
@@ -390,7 +460,9 @@ fmt.Println("v2 Encoder:\n" + string(buf.Bytes()))
|
|||||||
|
|
||||||
V1 always uses double quotes (`"`) around strings and keys that cannot be
|
V1 always uses double quotes (`"`) around strings and keys that cannot be
|
||||||
represented bare (unquoted). V2 uses single quotes instead by default (`'`),
|
represented bare (unquoted). V2 uses single quotes instead by default (`'`),
|
||||||
unless a character cannot be represented, then falls back to double quotes.
|
unless a character cannot be represented, then falls back to double quotes. As a
|
||||||
|
result of this change, `Encoder.QuoteMapKeys` has been removed, as it is not
|
||||||
|
useful anymore.
|
||||||
|
|
||||||
There is no way to make v2 encoder behave like v1.
|
There is no way to make v2 encoder behave like v1.
|
||||||
|
|
||||||
@@ -405,6 +477,76 @@ There is no way to make v2 encoder behave like v1.
|
|||||||
|
|
||||||
[tm]: https://golang.org/pkg/encoding/#TextMarshaler
|
[tm]: https://golang.org/pkg/encoding/#TextMarshaler
|
||||||
|
|
||||||
|
#### `Encoder.CompactComments` has been removed
|
||||||
|
|
||||||
|
Emitting compact comments is now the default behavior of go-toml. This option
|
||||||
|
is not necessary anymore.
|
||||||
|
|
||||||
|
#### Struct tags have been merged
|
||||||
|
|
||||||
|
V1 used to provide multiple struct tags: `comment`, `commented`, `multiline`,
|
||||||
|
`toml`, and `omitempty`. To behave more like the standard library, v2 has merged
|
||||||
|
`toml`, `multiline`, and `omitempty`. For example:
|
||||||
|
|
||||||
|
```go
|
||||||
|
type doc struct {
|
||||||
|
// v1
|
||||||
|
F string `toml:"field" multiline:"true" omitempty:"true"`
|
||||||
|
// v2
|
||||||
|
F string `toml:"field,multiline,omitempty"`
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Has a result, the `Encoder.SetTag*` methods have been removed, as there is just
|
||||||
|
one tag now.
|
||||||
|
|
||||||
|
|
||||||
|
#### `commented` tag has been removed
|
||||||
|
|
||||||
|
There is no replacement for the `commented` tag. This feature would be better
|
||||||
|
suited in a proper document model for go-toml v2, which has been [cut from
|
||||||
|
scope][nodoc] at the moment.
|
||||||
|
|
||||||
|
#### `Encoder.ArraysWithOneElementPerLine` has been renamed
|
||||||
|
|
||||||
|
The new name is `Encoder.SetArraysMultiline`. The behavior should be the same.
|
||||||
|
|
||||||
|
#### `Encoder.Indentation` has been renamed
|
||||||
|
|
||||||
|
The new name is `Encoder.SetIndentSymbol`. The behavior should be the same.
|
||||||
|
|
||||||
|
|
||||||
|
#### Embedded structs behave like stdlib
|
||||||
|
|
||||||
|
V1 defaults to merging embedded struct fields into the embedding struct. This
|
||||||
|
behavior was unexpected because it does not follow the standard library. To
|
||||||
|
avoid breaking backward compatibility, the `Encoder.PromoteAnonymous` method was
|
||||||
|
added to make the encoder behave correctly. Given backward compatibility is not
|
||||||
|
a problem anymore, v2 does the right thing by default: it follows the behavior
|
||||||
|
of `encoding/json`. `Encoder.PromoteAnonymous` has been removed.
|
||||||
|
|
||||||
|
[nodoc]: https://github.com/pelletier/go-toml/discussions/506#discussioncomment-1526038
|
||||||
|
|
||||||
|
### `query`
|
||||||
|
|
||||||
|
go-toml v1 provided the [`go-toml/query`][query] package. It allowed to run
|
||||||
|
JSONPath-style queries on TOML files. This feature is not available in v2. For a
|
||||||
|
replacement, check out [dasel][dasel].
|
||||||
|
|
||||||
|
This package has been removed because it was essentially not supported anymore
|
||||||
|
(last commit May 2020), increased the complexity of the code base, and more
|
||||||
|
complete solutions exist out there.
|
||||||
|
|
||||||
|
[query]: https://github.com/pelletier/go-toml/tree/f99d6bbca119636aeafcf351ee52b3d202782627/query
|
||||||
|
[dasel]: https://github.com/TomWright/dasel
|
||||||
|
|
||||||
|
## Versioning
|
||||||
|
|
||||||
|
Go-toml follows [Semantic Versioning](http://semver.org/). The supported version
|
||||||
|
of [TOML](https://github.com/toml-lang/toml) is indicated at the beginning of
|
||||||
|
this document. The last two major versions of Go are supported
|
||||||
|
(see [Go Release Policy](https://golang.org/doc/devel/release.html#policy)).
|
||||||
|
|
||||||
## License
|
## License
|
||||||
|
|
||||||
The MIT License (MIT). Read [LICENSE](LICENSE).
|
The MIT License (MIT). Read [LICENSE](LICENSE).
|
||||||
|
|||||||
+19
@@ -0,0 +1,19 @@
|
|||||||
|
# Security Policy
|
||||||
|
|
||||||
|
## Supported Versions
|
||||||
|
|
||||||
|
Use this section to tell people about which versions of your project are
|
||||||
|
currently being supported with security updates.
|
||||||
|
|
||||||
|
| Version | Supported |
|
||||||
|
| ---------- | ------------------ |
|
||||||
|
| Latest 2.x | :white_check_mark: |
|
||||||
|
| All 1.x | :x: |
|
||||||
|
| All 0.x | :x: |
|
||||||
|
|
||||||
|
## Reporting a Vulnerability
|
||||||
|
|
||||||
|
Email a vulnerability report to `security@pelletier.codes`. Make sure to include
|
||||||
|
as many details as possible to reproduce the vulnerability. This is a
|
||||||
|
side-project: I will try to get back to you as quickly as possible, time
|
||||||
|
permitting in my personal life. Providing a working patch helps very much!
|
||||||
@@ -0,0 +1,71 @@
|
|||||||
|
package toml
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
var valid10Ascii = []byte("1234567890")
|
||||||
|
var valid10Utf8 = []byte("日本語a")
|
||||||
|
var valid1kUtf8 = bytes.Repeat([]byte("0123456789日本語日本語日本語日abcdefghijklmnopqrstuvwx"), 16)
|
||||||
|
var valid1MUtf8 = bytes.Repeat(valid1kUtf8, 1024)
|
||||||
|
var valid1kAscii = bytes.Repeat([]byte("012345678998jhjklasDJKLAAdjdfjsdklfjdslkabcdefghijklmnopqrstuvwx"), 16)
|
||||||
|
var valid1MAscii = bytes.Repeat(valid1kAscii, 1024)
|
||||||
|
|
||||||
|
func BenchmarkScanComments(b *testing.B) {
|
||||||
|
wrap := func(x []byte) []byte {
|
||||||
|
return []byte("# " + string(x) + "\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
inputs := map[string][]byte{
|
||||||
|
"10Valid": wrap(valid10Ascii),
|
||||||
|
"1kValid": wrap(valid1kAscii),
|
||||||
|
"1MValid": wrap(valid1MAscii),
|
||||||
|
"10ValidUtf8": wrap(valid10Utf8),
|
||||||
|
"1kValidUtf8": wrap(valid1kUtf8),
|
||||||
|
"1MValidUtf8": wrap(valid1MUtf8),
|
||||||
|
}
|
||||||
|
|
||||||
|
for name, input := range inputs {
|
||||||
|
b.Run(name, func(b *testing.B) {
|
||||||
|
b.SetBytes(int64(len(input)))
|
||||||
|
b.ReportAllocs()
|
||||||
|
b.ResetTimer()
|
||||||
|
|
||||||
|
for i := 0; i < b.N; i++ {
|
||||||
|
scanComment(input)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func BenchmarkParseLiteralStringValid(b *testing.B) {
|
||||||
|
wrap := func(x []byte) []byte {
|
||||||
|
return []byte("'" + string(x) + "'")
|
||||||
|
}
|
||||||
|
|
||||||
|
inputs := map[string][]byte{
|
||||||
|
"10Valid": wrap(valid10Ascii),
|
||||||
|
"1kValid": wrap(valid1kAscii),
|
||||||
|
"1MValid": wrap(valid1MAscii),
|
||||||
|
"10ValidUtf8": wrap(valid10Utf8),
|
||||||
|
"1kValidUtf8": wrap(valid1kUtf8),
|
||||||
|
"1MValidUtf8": wrap(valid1MUtf8),
|
||||||
|
}
|
||||||
|
|
||||||
|
for name, input := range inputs {
|
||||||
|
b.Run(name, func(b *testing.B) {
|
||||||
|
p := parser{}
|
||||||
|
b.SetBytes(int64(len(input)))
|
||||||
|
b.ReportAllocs()
|
||||||
|
b.ResetTimer()
|
||||||
|
|
||||||
|
for i := 0; i < b.N; i++ {
|
||||||
|
_, _, _, err := p.parseLiteralString(input)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -76,7 +76,8 @@ cover() {
|
|||||||
fi
|
fi
|
||||||
|
|
||||||
pushd "$dir"
|
pushd "$dir"
|
||||||
go test -covermode=atomic -coverprofile=coverage.out ./...
|
go test -covermode=atomic -coverpkg=./... -coverprofile=coverage.out.tmp ./...
|
||||||
|
cat coverage.out.tmp | grep -v testsuite | grep -v tomltestgen | grep -v gotoml-test-decoder > coverage.out
|
||||||
go tool cover -func=coverage.out
|
go tool cover -func=coverage.out
|
||||||
popd
|
popd
|
||||||
|
|
||||||
@@ -103,16 +104,23 @@ coverage() {
|
|||||||
|
|
||||||
echo ""
|
echo ""
|
||||||
|
|
||||||
target_pct="$(cat ${target_out} |sed -E 's/.*total.*\t([0-9.]+)%/\1/;t;d')"
|
target_pct="$(tail -n2 ${target_out} | head -n1 | sed -E 's/.*total.*\t([0-9.]+)%.*/\1/')"
|
||||||
head_pct="$(cat ${head_out} |sed -E 's/.*total.*\t([0-9.]+)%/\1/;t;d')"
|
head_pct="$(tail -n2 ${head_out} | head -n1 | sed -E 's/.*total.*\t([0-9.]+)%/\1/')"
|
||||||
echo "Results: ${target} ${target_pct}% HEAD ${head_pct}%"
|
echo "Results: ${target} ${target_pct}% HEAD ${head_pct}%"
|
||||||
|
|
||||||
delta_pct=$(echo "$head_pct - $target_pct" | bc -l)
|
delta_pct=$(echo "$head_pct - $target_pct" | bc -l)
|
||||||
echo "Delta: ${delta_pct}"
|
echo "Delta: ${delta_pct}"
|
||||||
|
|
||||||
if [[ $delta_pct = \-* ]]; then
|
if [[ $delta_pct = \-* ]]; then
|
||||||
echo "Regression!";
|
echo "Regression!";
|
||||||
return 1
|
|
||||||
|
target_diff="${output_dir}/target.diff.txt"
|
||||||
|
head_diff="${output_dir}/head.diff.txt"
|
||||||
|
cat "${target_out}" | grep -E '^github.com/pelletier/go-toml' | tr -s "\t " | cut -f 2,3 | sort > "${target_diff}"
|
||||||
|
cat "${head_out}" | grep -E '^github.com/pelletier/go-toml' | tr -s "\t " | cut -f 2,3 | sort > "${head_diff}"
|
||||||
|
|
||||||
|
diff --side-by-side --suppress-common-lines "${target_diff}" "${head_diff}"
|
||||||
|
return 1
|
||||||
fi
|
fi
|
||||||
return 0
|
return 0
|
||||||
;;
|
;;
|
||||||
@@ -140,12 +148,10 @@ bench() {
|
|||||||
if [ "${replace}" != "" ]; then
|
if [ "${replace}" != "" ]; then
|
||||||
find ./benchmark/ -iname '*.go' -exec sed -i -E "s|github.com/pelletier/go-toml/v2|${replace}|g" {} \;
|
find ./benchmark/ -iname '*.go' -exec sed -i -E "s|github.com/pelletier/go-toml/v2|${replace}|g" {} \;
|
||||||
go get "${replace}"
|
go get "${replace}"
|
||||||
# hack: remove canada.toml.gz because it is not supported by
|
|
||||||
# burntsushi, and replace is only used for benchmark -a
|
|
||||||
rm -f benchmark/testdata/canada.toml.gz
|
|
||||||
fi
|
fi
|
||||||
|
|
||||||
go test -bench=. -count=10 ./... | tee "${out}"
|
export GOMAXPROCS=2
|
||||||
|
nice -n -19 taskset --cpu-list 0,1 go test '-bench=^Benchmark(Un)?[mM]arshal' -count=5 -run=Nothing ./... | tee "${out}"
|
||||||
popd
|
popd
|
||||||
|
|
||||||
if [ "${branch}" != "HEAD" ]; then
|
if [ "${branch}" != "HEAD" ]; then
|
||||||
|
|||||||
@@ -0,0 +1,30 @@
|
|||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"flag"
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"path"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2/internal/testsuite"
|
||||||
|
)
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
log.SetFlags(0)
|
||||||
|
flag.Usage = usage
|
||||||
|
flag.Parse()
|
||||||
|
if flag.NArg() != 0 {
|
||||||
|
flag.Usage()
|
||||||
|
}
|
||||||
|
|
||||||
|
err := testsuite.DecodeStdin()
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func usage() {
|
||||||
|
log.Printf("Usage: %s < toml-file\n", path.Base(os.Args[0]))
|
||||||
|
flag.PrintDefaults()
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
@@ -0,0 +1,55 @@
|
|||||||
|
// Package jsontoml is a program that converts JSON to TOML.
|
||||||
|
//
|
||||||
|
// # Usage
|
||||||
|
//
|
||||||
|
// Reading from stdin:
|
||||||
|
//
|
||||||
|
// cat file.json | jsontoml > file.toml
|
||||||
|
//
|
||||||
|
// Reading from a file:
|
||||||
|
//
|
||||||
|
// jsontoml file.json > file.toml
|
||||||
|
//
|
||||||
|
// # Installation
|
||||||
|
//
|
||||||
|
// Using Go:
|
||||||
|
//
|
||||||
|
// go install github.com/pelletier/go-toml/v2/cmd/jsontoml@latest
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"io"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
"github.com/pelletier/go-toml/v2/internal/cli"
|
||||||
|
)
|
||||||
|
|
||||||
|
const usage = `jsontoml can be used in two ways:
|
||||||
|
Reading from stdin:
|
||||||
|
cat file.json | jsontoml > file.toml
|
||||||
|
|
||||||
|
Reading from a file:
|
||||||
|
jsontoml file.json > file.toml
|
||||||
|
`
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
p := cli.Program{
|
||||||
|
Usage: usage,
|
||||||
|
Fn: convert,
|
||||||
|
}
|
||||||
|
p.Execute()
|
||||||
|
}
|
||||||
|
|
||||||
|
func convert(r io.Reader, w io.Writer) error {
|
||||||
|
var v interface{}
|
||||||
|
|
||||||
|
d := json.NewDecoder(r)
|
||||||
|
err := d.Decode(&v)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
e := toml.NewEncoder(w)
|
||||||
|
return e.Encode(v)
|
||||||
|
}
|
||||||
@@ -0,0 +1,48 @@
|
|||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestConvert(t *testing.T) {
|
||||||
|
examples := []struct {
|
||||||
|
name string
|
||||||
|
input string
|
||||||
|
expected string
|
||||||
|
errors bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "valid json",
|
||||||
|
input: `
|
||||||
|
{
|
||||||
|
"mytoml": {
|
||||||
|
"a": 42
|
||||||
|
}
|
||||||
|
}`,
|
||||||
|
expected: `[mytoml]
|
||||||
|
a = 42.0
|
||||||
|
`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "invalid json",
|
||||||
|
input: `{ foo`,
|
||||||
|
errors: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, e := range examples {
|
||||||
|
b := new(bytes.Buffer)
|
||||||
|
err := convert(strings.NewReader(e.input), b)
|
||||||
|
if e.errors {
|
||||||
|
require.Error(t, err)
|
||||||
|
} else {
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.Equal(t, e.expected, b.String())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -0,0 +1,63 @@
|
|||||||
|
// Package tomljson is a program that converts TOML to JSON.
|
||||||
|
//
|
||||||
|
// # Usage
|
||||||
|
//
|
||||||
|
// Reading from stdin:
|
||||||
|
//
|
||||||
|
// cat file.toml | tomljson > file.json
|
||||||
|
//
|
||||||
|
// Reading from a file:
|
||||||
|
//
|
||||||
|
// tomljson file.toml > file.json
|
||||||
|
//
|
||||||
|
// # Installation
|
||||||
|
//
|
||||||
|
// Using Go:
|
||||||
|
//
|
||||||
|
// go install github.com/pelletier/go-toml/v2/cmd/tomljson@latest
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
"github.com/pelletier/go-toml/v2/internal/cli"
|
||||||
|
)
|
||||||
|
|
||||||
|
const usage = `tomljson can be used in two ways:
|
||||||
|
Reading from stdin:
|
||||||
|
cat file.toml | tomljson > file.json
|
||||||
|
|
||||||
|
Reading from a file:
|
||||||
|
tomljson file.toml > file.json
|
||||||
|
`
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
p := cli.Program{
|
||||||
|
Usage: usage,
|
||||||
|
Fn: convert,
|
||||||
|
}
|
||||||
|
p.Execute()
|
||||||
|
}
|
||||||
|
|
||||||
|
func convert(r io.Reader, w io.Writer) error {
|
||||||
|
var v interface{}
|
||||||
|
|
||||||
|
d := toml.NewDecoder(r)
|
||||||
|
err := d.Decode(&v)
|
||||||
|
if err != nil {
|
||||||
|
var derr *toml.DecodeError
|
||||||
|
if errors.As(err, &derr) {
|
||||||
|
row, col := derr.Position()
|
||||||
|
return fmt.Errorf("%s\nerror occurred at row %d column %d", derr.String(), row, col)
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
e := json.NewEncoder(w)
|
||||||
|
e.SetIndent("", " ")
|
||||||
|
return e.Encode(v)
|
||||||
|
}
|
||||||
@@ -0,0 +1,61 @@
|
|||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestConvert(t *testing.T) {
|
||||||
|
examples := []struct {
|
||||||
|
name string
|
||||||
|
input io.Reader
|
||||||
|
expected string
|
||||||
|
errors bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "valid toml",
|
||||||
|
input: strings.NewReader(`
|
||||||
|
[mytoml]
|
||||||
|
a = 42`),
|
||||||
|
expected: `{
|
||||||
|
"mytoml": {
|
||||||
|
"a": 42
|
||||||
|
}
|
||||||
|
}
|
||||||
|
`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "invalid toml",
|
||||||
|
input: strings.NewReader(`bad = []]`),
|
||||||
|
errors: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "bad reader",
|
||||||
|
input: &badReader{},
|
||||||
|
errors: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, e := range examples {
|
||||||
|
b := new(bytes.Buffer)
|
||||||
|
err := convert(e.input, b)
|
||||||
|
if e.errors {
|
||||||
|
require.Error(t, err)
|
||||||
|
} else {
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.Equal(t, e.expected, b.String())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type badReader struct{}
|
||||||
|
|
||||||
|
func (r *badReader) Read([]byte) (int, error) {
|
||||||
|
return 0, fmt.Errorf("reader failed on purpose")
|
||||||
|
}
|
||||||
@@ -0,0 +1,58 @@
|
|||||||
|
// Package tomll is a linter program for TOML.
|
||||||
|
//
|
||||||
|
// # Usage
|
||||||
|
//
|
||||||
|
// Reading from stdin, writing to stdout:
|
||||||
|
//
|
||||||
|
// cat file.toml | tomll
|
||||||
|
//
|
||||||
|
// Reading and updating a list of files in place:
|
||||||
|
//
|
||||||
|
// tomll a.toml b.toml c.toml
|
||||||
|
//
|
||||||
|
// # Installation
|
||||||
|
//
|
||||||
|
// Using Go:
|
||||||
|
//
|
||||||
|
// go install github.com/pelletier/go-toml/v2/cmd/tomll@latest
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"io"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
"github.com/pelletier/go-toml/v2/internal/cli"
|
||||||
|
)
|
||||||
|
|
||||||
|
const usage = `tomll can be used in two ways:
|
||||||
|
|
||||||
|
Reading from stdin, writing to stdout:
|
||||||
|
cat file.toml | tomll > file.toml
|
||||||
|
|
||||||
|
Reading and updating a list of files in place:
|
||||||
|
tomll a.toml b.toml c.toml
|
||||||
|
|
||||||
|
When given a list of files, tomll will modify all files in place without asking.
|
||||||
|
`
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
p := cli.Program{
|
||||||
|
Usage: usage,
|
||||||
|
Fn: convert,
|
||||||
|
Inplace: true,
|
||||||
|
}
|
||||||
|
p.Execute()
|
||||||
|
}
|
||||||
|
|
||||||
|
func convert(r io.Reader, w io.Writer) error {
|
||||||
|
var v interface{}
|
||||||
|
|
||||||
|
d := toml.NewDecoder(r)
|
||||||
|
err := d.Decode(&v)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
e := toml.NewEncoder(w)
|
||||||
|
return e.Encode(v)
|
||||||
|
}
|
||||||
@@ -0,0 +1,45 @@
|
|||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestConvert(t *testing.T) {
|
||||||
|
examples := []struct {
|
||||||
|
name string
|
||||||
|
input string
|
||||||
|
expected string
|
||||||
|
errors bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "valid toml",
|
||||||
|
input: `
|
||||||
|
mytoml.a = 42.0
|
||||||
|
`,
|
||||||
|
expected: `[mytoml]
|
||||||
|
a = 42.0
|
||||||
|
`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "invalid toml",
|
||||||
|
input: `[what`,
|
||||||
|
errors: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, e := range examples {
|
||||||
|
b := new(bytes.Buffer)
|
||||||
|
err := convert(strings.NewReader(e.input), b)
|
||||||
|
if e.errors {
|
||||||
|
require.Error(t, err)
|
||||||
|
} else {
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.Equal(t, e.expected, b.String())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -0,0 +1,223 @@
|
|||||||
|
// tomltestgen retrieves a given version of the language-agnostic TOML test suite in
|
||||||
|
// https://github.com/BurntSushi/toml-test and generates go-toml unit tests.
|
||||||
|
//
|
||||||
|
// Within the go-toml package, run `go generate`. Otherwise, use:
|
||||||
|
//
|
||||||
|
// go run github.com/pelletier/go-toml/cmd/tomltestgen -o toml_testgen_test.go
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"archive/zip"
|
||||||
|
"bytes"
|
||||||
|
"flag"
|
||||||
|
"fmt"
|
||||||
|
"go/format"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"log"
|
||||||
|
"net/http"
|
||||||
|
"os"
|
||||||
|
"regexp"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"text/template"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
type invalid struct {
|
||||||
|
Name string
|
||||||
|
Input string
|
||||||
|
}
|
||||||
|
|
||||||
|
type valid struct {
|
||||||
|
Name string
|
||||||
|
Input string
|
||||||
|
JsonRef string
|
||||||
|
}
|
||||||
|
|
||||||
|
type testsCollection struct {
|
||||||
|
Ref string
|
||||||
|
Timestamp string
|
||||||
|
Invalid []invalid
|
||||||
|
Valid []valid
|
||||||
|
Count int
|
||||||
|
}
|
||||||
|
|
||||||
|
const srcTemplate = "// Generated by tomltestgen for toml-test ref {{.Ref}} on {{.Timestamp}}\n" +
|
||||||
|
"package toml_test\n" +
|
||||||
|
" import (\n" +
|
||||||
|
" \"testing\"\n" +
|
||||||
|
")\n" +
|
||||||
|
|
||||||
|
"{{range .Invalid}}\n" +
|
||||||
|
"func TestTOMLTest_Invalid_{{.Name}}(t *testing.T) {\n" +
|
||||||
|
" input := {{.Input|gostr}}\n" +
|
||||||
|
" testgenInvalid(t, input)\n" +
|
||||||
|
"}\n" +
|
||||||
|
"{{end}}\n" +
|
||||||
|
"\n" +
|
||||||
|
"{{range .Valid}}\n" +
|
||||||
|
"func TestTOMLTest_Valid_{{.Name}}(t *testing.T) {\n" +
|
||||||
|
" input := {{.Input|gostr}}\n" +
|
||||||
|
" jsonRef := {{.JsonRef|gostr}}\n" +
|
||||||
|
" testgenValid(t, input, jsonRef)\n" +
|
||||||
|
"}\n" +
|
||||||
|
"{{end}}\n"
|
||||||
|
|
||||||
|
func downloadTmpFile(url string) string {
|
||||||
|
log.Println("starting to download file from", url)
|
||||||
|
resp, err := http.Get(url)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
|
||||||
|
tmpfile, err := ioutil.TempFile("", "toml-test-*.zip")
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
defer tmpfile.Close()
|
||||||
|
|
||||||
|
copiedLen, err := io.Copy(tmpfile, resp.Body)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
if resp.ContentLength > 0 && copiedLen != resp.ContentLength {
|
||||||
|
panic(fmt.Errorf("copied %d bytes, request body had %d", copiedLen, resp.ContentLength))
|
||||||
|
}
|
||||||
|
return tmpfile.Name()
|
||||||
|
}
|
||||||
|
|
||||||
|
func kebabToCamel(kebab string) string {
|
||||||
|
camel := ""
|
||||||
|
nextUpper := true
|
||||||
|
for _, c := range kebab {
|
||||||
|
if nextUpper {
|
||||||
|
camel += strings.ToUpper(string(c))
|
||||||
|
nextUpper = false
|
||||||
|
} else if c == '-' {
|
||||||
|
nextUpper = true
|
||||||
|
} else if c == '/' {
|
||||||
|
nextUpper = true
|
||||||
|
camel += "_"
|
||||||
|
} else {
|
||||||
|
camel += string(c)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return camel
|
||||||
|
}
|
||||||
|
|
||||||
|
func readFileFromZip(f *zip.File) string {
|
||||||
|
reader, err := f.Open()
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
defer reader.Close()
|
||||||
|
bytes, err := ioutil.ReadAll(reader)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
return string(bytes)
|
||||||
|
}
|
||||||
|
|
||||||
|
func templateGoStr(input string) string {
|
||||||
|
return strconv.Quote(input)
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
ref = flag.String("r", "master", "git reference")
|
||||||
|
out = flag.String("o", "", "output file")
|
||||||
|
)
|
||||||
|
|
||||||
|
func usage() {
|
||||||
|
_, _ = fmt.Fprintf(os.Stderr, "usage: tomltestgen [flags]\n")
|
||||||
|
flag.PrintDefaults()
|
||||||
|
}
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
flag.Usage = usage
|
||||||
|
flag.Parse()
|
||||||
|
|
||||||
|
url := "https://codeload.github.com/BurntSushi/toml-test/zip/" + *ref
|
||||||
|
resultFile := downloadTmpFile(url)
|
||||||
|
defer os.Remove(resultFile)
|
||||||
|
log.Println("file written to", resultFile)
|
||||||
|
|
||||||
|
zipReader, err := zip.OpenReader(resultFile)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
defer zipReader.Close()
|
||||||
|
|
||||||
|
collection := testsCollection{
|
||||||
|
Ref: *ref,
|
||||||
|
Timestamp: time.Now().Format(time.RFC3339),
|
||||||
|
}
|
||||||
|
|
||||||
|
zipFilesMap := map[string]*zip.File{}
|
||||||
|
|
||||||
|
for _, f := range zipReader.File {
|
||||||
|
zipFilesMap[f.Name] = f
|
||||||
|
}
|
||||||
|
|
||||||
|
testFileRegexp := regexp.MustCompile(`([^/]+/tests/(valid|invalid)/(.+))\.(toml)`)
|
||||||
|
for _, f := range zipReader.File {
|
||||||
|
groups := testFileRegexp.FindStringSubmatch(f.Name)
|
||||||
|
if len(groups) > 0 {
|
||||||
|
name := kebabToCamel(groups[3])
|
||||||
|
testType := groups[2]
|
||||||
|
|
||||||
|
log.Printf("> [%s] %s\n", testType, name)
|
||||||
|
|
||||||
|
tomlContent := readFileFromZip(f)
|
||||||
|
|
||||||
|
switch testType {
|
||||||
|
case "invalid":
|
||||||
|
collection.Invalid = append(collection.Invalid, invalid{
|
||||||
|
Name: name,
|
||||||
|
Input: tomlContent,
|
||||||
|
})
|
||||||
|
collection.Count++
|
||||||
|
case "valid":
|
||||||
|
baseFilePath := groups[1]
|
||||||
|
jsonFilePath := baseFilePath + ".json"
|
||||||
|
jsonContent := readFileFromZip(zipFilesMap[jsonFilePath])
|
||||||
|
|
||||||
|
collection.Valid = append(collection.Valid, valid{
|
||||||
|
Name: name,
|
||||||
|
Input: tomlContent,
|
||||||
|
JsonRef: jsonContent,
|
||||||
|
})
|
||||||
|
collection.Count++
|
||||||
|
default:
|
||||||
|
panic(fmt.Sprintf("unknown test type: %s", testType))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Printf("Collected %d tests from toml-test\n", collection.Count)
|
||||||
|
|
||||||
|
funcMap := template.FuncMap{
|
||||||
|
"gostr": templateGoStr,
|
||||||
|
}
|
||||||
|
t := template.Must(template.New("src").Funcs(funcMap).Parse(srcTemplate))
|
||||||
|
buf := new(bytes.Buffer)
|
||||||
|
err = t.Execute(buf, collection)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
outputBytes, err := format.Source(buf.Bytes())
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if *out == "" {
|
||||||
|
fmt.Println(string(outputBytes))
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
err = os.WriteFile(*out, outputBytes, 0644)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -35,26 +35,42 @@ func parseLocalDate(b []byte) (LocalDate, error) {
|
|||||||
return date, newDecodeError(b, "dates are expected to have the format YYYY-MM-DD")
|
return date, newDecodeError(b, "dates are expected to have the format YYYY-MM-DD")
|
||||||
}
|
}
|
||||||
|
|
||||||
date.Year = parseDecimalDigits(b[0:4])
|
var err error
|
||||||
|
|
||||||
v := parseDecimalDigits(b[5:7])
|
date.Year, err = parseDecimalDigits(b[0:4])
|
||||||
|
if err != nil {
|
||||||
|
return LocalDate{}, err
|
||||||
|
}
|
||||||
|
|
||||||
date.Month = time.Month(v)
|
date.Month, err = parseDecimalDigits(b[5:7])
|
||||||
|
if err != nil {
|
||||||
|
return LocalDate{}, err
|
||||||
|
}
|
||||||
|
|
||||||
date.Day = parseDecimalDigits(b[8:10])
|
date.Day, err = parseDecimalDigits(b[8:10])
|
||||||
|
if err != nil {
|
||||||
|
return LocalDate{}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if !isValidDate(date.Year, date.Month, date.Day) {
|
||||||
|
return LocalDate{}, newDecodeError(b, "impossible date")
|
||||||
|
}
|
||||||
|
|
||||||
return date, nil
|
return date, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func parseDecimalDigits(b []byte) int {
|
func parseDecimalDigits(b []byte) (int, error) {
|
||||||
v := 0
|
v := 0
|
||||||
|
|
||||||
for _, c := range b {
|
for i, c := range b {
|
||||||
|
if c < '0' || c > '9' {
|
||||||
|
return 0, newDecodeError(b[i:i+1], "expected digit (0-9)")
|
||||||
|
}
|
||||||
v *= 10
|
v *= 10
|
||||||
v += int(c - '0')
|
v += int(c - '0')
|
||||||
}
|
}
|
||||||
|
|
||||||
return v
|
return v, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func parseDateTime(b []byte) (time.Time, error) {
|
func parseDateTime(b []byte) (time.Time, error) {
|
||||||
@@ -75,7 +91,7 @@ func parseDateTime(b []byte) (time.Time, error) {
|
|||||||
panic("date time should have a timezone")
|
panic("date time should have a timezone")
|
||||||
}
|
}
|
||||||
|
|
||||||
if b[0] == 'Z' {
|
if b[0] == 'Z' || b[0] == 'z' {
|
||||||
b = b[1:]
|
b = b[1:]
|
||||||
zone = time.UTC
|
zone = time.UTC
|
||||||
} else {
|
} else {
|
||||||
@@ -83,15 +99,42 @@ func parseDateTime(b []byte) (time.Time, error) {
|
|||||||
if len(b) != dateTimeByteLen {
|
if len(b) != dateTimeByteLen {
|
||||||
return time.Time{}, newDecodeError(b, "invalid date-time timezone")
|
return time.Time{}, newDecodeError(b, "invalid date-time timezone")
|
||||||
}
|
}
|
||||||
direction := 1
|
var direction int
|
||||||
if b[0] == '-' {
|
switch b[0] {
|
||||||
|
case '-':
|
||||||
direction = -1
|
direction = -1
|
||||||
|
case '+':
|
||||||
|
direction = +1
|
||||||
|
default:
|
||||||
|
return time.Time{}, newDecodeError(b[:1], "invalid timezone offset character")
|
||||||
|
}
|
||||||
|
|
||||||
|
if b[3] != ':' {
|
||||||
|
return time.Time{}, newDecodeError(b[3:4], "expected a : separator")
|
||||||
|
}
|
||||||
|
|
||||||
|
hours, err := parseDecimalDigits(b[1:3])
|
||||||
|
if err != nil {
|
||||||
|
return time.Time{}, err
|
||||||
|
}
|
||||||
|
if hours > 23 {
|
||||||
|
return time.Time{}, newDecodeError(b[:1], "invalid timezone offset hours")
|
||||||
|
}
|
||||||
|
|
||||||
|
minutes, err := parseDecimalDigits(b[4:6])
|
||||||
|
if err != nil {
|
||||||
|
return time.Time{}, err
|
||||||
|
}
|
||||||
|
if minutes > 59 {
|
||||||
|
return time.Time{}, newDecodeError(b[:1], "invalid timezone offset minutes")
|
||||||
}
|
}
|
||||||
|
|
||||||
hours := digitsToInt(b[1:3])
|
|
||||||
minutes := digitsToInt(b[4:6])
|
|
||||||
seconds := direction * (hours*3600 + minutes*60)
|
seconds := direction * (hours*3600 + minutes*60)
|
||||||
zone = time.FixedZone("", seconds)
|
if seconds == 0 {
|
||||||
|
zone = time.UTC
|
||||||
|
} else {
|
||||||
|
zone = time.FixedZone("", seconds)
|
||||||
|
}
|
||||||
b = b[dateTimeByteLen:]
|
b = b[dateTimeByteLen:]
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -100,13 +143,13 @@ func parseDateTime(b []byte) (time.Time, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
t := time.Date(
|
t := time.Date(
|
||||||
dt.Date.Year,
|
dt.Year,
|
||||||
dt.Date.Month,
|
time.Month(dt.Month),
|
||||||
dt.Date.Day,
|
dt.Day,
|
||||||
dt.Time.Hour,
|
dt.Hour,
|
||||||
dt.Time.Minute,
|
dt.Minute,
|
||||||
dt.Time.Second,
|
dt.Second,
|
||||||
dt.Time.Nanosecond,
|
dt.Nanosecond,
|
||||||
zone)
|
zone)
|
||||||
|
|
||||||
return t, nil
|
return t, nil
|
||||||
@@ -124,10 +167,10 @@ func parseLocalDateTime(b []byte) (LocalDateTime, []byte, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return dt, nil, err
|
return dt, nil, err
|
||||||
}
|
}
|
||||||
dt.Date = date
|
dt.LocalDate = date
|
||||||
|
|
||||||
sep := b[10]
|
sep := b[10]
|
||||||
if sep != 'T' && sep != ' ' {
|
if sep != 'T' && sep != ' ' && sep != 't' {
|
||||||
return dt, nil, newDecodeError(b[10:11], "datetime separator is expected to be T or a space")
|
return dt, nil, newDecodeError(b[10:11], "datetime separator is expected to be T or a space")
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -135,7 +178,7 @@ func parseLocalDateTime(b []byte) (LocalDateTime, []byte, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return dt, nil, err
|
return dt, nil, err
|
||||||
}
|
}
|
||||||
dt.Time = t
|
dt.LocalTime = t
|
||||||
|
|
||||||
return dt, rest, nil
|
return dt, rest, nil
|
||||||
}
|
}
|
||||||
@@ -149,53 +192,89 @@ func parseLocalTime(b []byte) (LocalTime, []byte, error) {
|
|||||||
t LocalTime
|
t LocalTime
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// check if b matches to have expected format HH:MM:SS[.NNNNNN]
|
||||||
const localTimeByteLen = 8
|
const localTimeByteLen = 8
|
||||||
if len(b) < localTimeByteLen {
|
if len(b) < localTimeByteLen {
|
||||||
return t, nil, newDecodeError(b, "times are expected to have the format HH:MM:SS[.NNNNNN]")
|
return t, nil, newDecodeError(b, "times are expected to have the format HH:MM:SS[.NNNNNN]")
|
||||||
}
|
}
|
||||||
|
|
||||||
t.Hour = parseDecimalDigits(b[0:2])
|
var err error
|
||||||
|
|
||||||
|
t.Hour, err = parseDecimalDigits(b[0:2])
|
||||||
|
if err != nil {
|
||||||
|
return t, nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if t.Hour > 23 {
|
||||||
|
return t, nil, newDecodeError(b[0:2], "hour cannot be greater 23")
|
||||||
|
}
|
||||||
if b[2] != ':' {
|
if b[2] != ':' {
|
||||||
return t, nil, newDecodeError(b[2:3], "expecting colon between hours and minutes")
|
return t, nil, newDecodeError(b[2:3], "expecting colon between hours and minutes")
|
||||||
}
|
}
|
||||||
|
|
||||||
t.Minute = parseDecimalDigits(b[3:5])
|
t.Minute, err = parseDecimalDigits(b[3:5])
|
||||||
|
if err != nil {
|
||||||
|
return t, nil, err
|
||||||
|
}
|
||||||
|
if t.Minute > 59 {
|
||||||
|
return t, nil, newDecodeError(b[3:5], "minutes cannot be greater 59")
|
||||||
|
}
|
||||||
if b[5] != ':' {
|
if b[5] != ':' {
|
||||||
return t, nil, newDecodeError(b[5:6], "expecting colon between minutes and seconds")
|
return t, nil, newDecodeError(b[5:6], "expecting colon between minutes and seconds")
|
||||||
}
|
}
|
||||||
|
|
||||||
t.Second = parseDecimalDigits(b[6:8])
|
t.Second, err = parseDecimalDigits(b[6:8])
|
||||||
|
if err != nil {
|
||||||
|
return t, nil, err
|
||||||
|
}
|
||||||
|
|
||||||
const minLengthWithFrac = 9
|
if t.Second > 60 {
|
||||||
if len(b) >= minLengthWithFrac && b[minLengthWithFrac-1] == '.' {
|
return t, nil, newDecodeError(b[6:8], "seconds cannot be greater 60")
|
||||||
|
}
|
||||||
|
|
||||||
|
b = b[8:]
|
||||||
|
|
||||||
|
if len(b) >= 1 && b[0] == '.' {
|
||||||
frac := 0
|
frac := 0
|
||||||
|
precision := 0
|
||||||
digits := 0
|
digits := 0
|
||||||
|
|
||||||
for i, c := range b[minLengthWithFrac:] {
|
for i, c := range b[1:] {
|
||||||
if !isDigit(c) {
|
if !isDigit(c) {
|
||||||
if i == 0 {
|
if i == 0 {
|
||||||
return t, nil, newDecodeError(b[i:i+1], "need at least one digit after fraction point")
|
return t, nil, newDecodeError(b[0:1], "need at least one digit after fraction point")
|
||||||
}
|
}
|
||||||
|
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
digits++
|
||||||
|
|
||||||
const maxFracPrecision = 9
|
const maxFracPrecision = 9
|
||||||
if i >= maxFracPrecision {
|
if i >= maxFracPrecision {
|
||||||
return t, nil, newDecodeError(b[i:i+1], "maximum precision for date time is nanosecond")
|
// go-toml allows decoding fractional seconds
|
||||||
|
// beyond the supported precision of 9
|
||||||
|
// digits. It truncates the fractional component
|
||||||
|
// to the supported precision and ignores the
|
||||||
|
// remaining digits.
|
||||||
|
//
|
||||||
|
// https://github.com/pelletier/go-toml/discussions/707
|
||||||
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
frac *= 10
|
frac *= 10
|
||||||
frac += int(c - '0')
|
frac += int(c - '0')
|
||||||
digits++
|
precision++
|
||||||
}
|
}
|
||||||
|
|
||||||
t.Nanosecond = frac * nspow[digits]
|
if precision == 0 {
|
||||||
|
return t, nil, newDecodeError(b[:1], "nanoseconds need at least one digit")
|
||||||
|
}
|
||||||
|
|
||||||
return t, b[9+digits:], nil
|
t.Nanosecond = frac * nspow[precision]
|
||||||
|
t.Precision = precision
|
||||||
|
|
||||||
|
return t, b[1+digits:], nil
|
||||||
}
|
}
|
||||||
|
return t, b, nil
|
||||||
return t, b[8:], nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
//nolint:cyclop
|
//nolint:cyclop
|
||||||
@@ -204,7 +283,7 @@ func parseFloat(b []byte) (float64, error) {
|
|||||||
return math.NaN(), nil
|
return math.NaN(), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
cleaned, err := checkAndRemoveUnderscores(b)
|
cleaned, err := checkAndRemoveUnderscoresFloats(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
@@ -217,6 +296,30 @@ func parseFloat(b []byte) (float64, error) {
|
|||||||
return 0, newDecodeError(b, "float cannot end with a dot")
|
return 0, newDecodeError(b, "float cannot end with a dot")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
dotAlreadySeen := false
|
||||||
|
for i, c := range cleaned {
|
||||||
|
if c == '.' {
|
||||||
|
if dotAlreadySeen {
|
||||||
|
return 0, newDecodeError(b[i:i+1], "float can have at most one decimal point")
|
||||||
|
}
|
||||||
|
if !isDigit(cleaned[i-1]) {
|
||||||
|
return 0, newDecodeError(b[i-1:i+1], "float decimal point must be preceded by a digit")
|
||||||
|
}
|
||||||
|
if !isDigit(cleaned[i+1]) {
|
||||||
|
return 0, newDecodeError(b[i:i+2], "float decimal point must be followed by a digit")
|
||||||
|
}
|
||||||
|
dotAlreadySeen = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
start := 0
|
||||||
|
if cleaned[0] == '+' || cleaned[0] == '-' {
|
||||||
|
start = 1
|
||||||
|
}
|
||||||
|
if cleaned[start] == '0' && isDigit(cleaned[start+1]) {
|
||||||
|
return 0, newDecodeError(b, "float integer part cannot have leading zeroes")
|
||||||
|
}
|
||||||
|
|
||||||
f, err := strconv.ParseFloat(string(cleaned), 64)
|
f, err := strconv.ParseFloat(string(cleaned), 64)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, newDecodeError(b, "unable to parse float: %w", err)
|
return 0, newDecodeError(b, "unable to parse float: %w", err)
|
||||||
@@ -226,7 +329,7 @@ func parseFloat(b []byte) (float64, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func parseIntHex(b []byte) (int64, error) {
|
func parseIntHex(b []byte) (int64, error) {
|
||||||
cleaned, err := checkAndRemoveUnderscores(b[2:])
|
cleaned, err := checkAndRemoveUnderscoresIntegers(b[2:])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
@@ -240,7 +343,7 @@ func parseIntHex(b []byte) (int64, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func parseIntOct(b []byte) (int64, error) {
|
func parseIntOct(b []byte) (int64, error) {
|
||||||
cleaned, err := checkAndRemoveUnderscores(b[2:])
|
cleaned, err := checkAndRemoveUnderscoresIntegers(b[2:])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
@@ -254,7 +357,7 @@ func parseIntOct(b []byte) (int64, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func parseIntBin(b []byte) (int64, error) {
|
func parseIntBin(b []byte) (int64, error) {
|
||||||
cleaned, err := checkAndRemoveUnderscores(b[2:])
|
cleaned, err := checkAndRemoveUnderscoresIntegers(b[2:])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
@@ -267,12 +370,26 @@ func parseIntBin(b []byte) (int64, error) {
|
|||||||
return i, nil
|
return i, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func isSign(b byte) bool {
|
||||||
|
return b == '+' || b == '-'
|
||||||
|
}
|
||||||
|
|
||||||
func parseIntDec(b []byte) (int64, error) {
|
func parseIntDec(b []byte) (int64, error) {
|
||||||
cleaned, err := checkAndRemoveUnderscores(b)
|
cleaned, err := checkAndRemoveUnderscoresIntegers(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
startIdx := 0
|
||||||
|
|
||||||
|
if isSign(cleaned[0]) {
|
||||||
|
startIdx++
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(cleaned) > startIdx+1 && cleaned[startIdx] == '0' {
|
||||||
|
return 0, newDecodeError(b, "leading zero not allowed on decimal number")
|
||||||
|
}
|
||||||
|
|
||||||
i, err := strconv.ParseInt(string(cleaned), 10, 64)
|
i, err := strconv.ParseInt(string(cleaned), 10, 64)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, newDecodeError(b, "couldn't parse decimal number: %w", err)
|
return 0, newDecodeError(b, "couldn't parse decimal number: %w", err)
|
||||||
@@ -281,9 +398,18 @@ func parseIntDec(b []byte) (int64, error) {
|
|||||||
return i, nil
|
return i, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func checkAndRemoveUnderscores(b []byte) ([]byte, error) {
|
func checkAndRemoveUnderscoresIntegers(b []byte) ([]byte, error) {
|
||||||
if b[0] == '_' {
|
start := 0
|
||||||
return nil, newDecodeError(b[0:1], "number cannot start with underscore")
|
if b[start] == '+' || b[start] == '-' {
|
||||||
|
start++
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(b) == start {
|
||||||
|
return b, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
if b[start] == '_' {
|
||||||
|
return nil, newDecodeError(b[start:start+1], "number cannot start with underscore")
|
||||||
}
|
}
|
||||||
|
|
||||||
if b[len(b)-1] == '_' {
|
if b[len(b)-1] == '_' {
|
||||||
@@ -320,3 +446,99 @@ func checkAndRemoveUnderscores(b []byte) ([]byte, error) {
|
|||||||
|
|
||||||
return cleaned, nil
|
return cleaned, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func checkAndRemoveUnderscoresFloats(b []byte) ([]byte, error) {
|
||||||
|
if b[0] == '_' {
|
||||||
|
return nil, newDecodeError(b[0:1], "number cannot start with underscore")
|
||||||
|
}
|
||||||
|
|
||||||
|
if b[len(b)-1] == '_' {
|
||||||
|
return nil, newDecodeError(b[len(b)-1:], "number cannot end with underscore")
|
||||||
|
}
|
||||||
|
|
||||||
|
// fast path
|
||||||
|
i := 0
|
||||||
|
for ; i < len(b); i++ {
|
||||||
|
if b[i] == '_' {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if i == len(b) {
|
||||||
|
return b, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
before := false
|
||||||
|
cleaned := make([]byte, 0, len(b))
|
||||||
|
|
||||||
|
for i := 0; i < len(b); i++ {
|
||||||
|
c := b[i]
|
||||||
|
|
||||||
|
switch c {
|
||||||
|
case '_':
|
||||||
|
if !before {
|
||||||
|
return nil, newDecodeError(b[i-1:i+1], "number must have at least one digit between underscores")
|
||||||
|
}
|
||||||
|
if i < len(b)-1 && (b[i+1] == 'e' || b[i+1] == 'E') {
|
||||||
|
return nil, newDecodeError(b[i+1:i+2], "cannot have underscore before exponent")
|
||||||
|
}
|
||||||
|
before = false
|
||||||
|
case '+', '-':
|
||||||
|
// signed exponents
|
||||||
|
cleaned = append(cleaned, c)
|
||||||
|
before = false
|
||||||
|
case 'e', 'E':
|
||||||
|
if i < len(b)-1 && b[i+1] == '_' {
|
||||||
|
return nil, newDecodeError(b[i+1:i+2], "cannot have underscore after exponent")
|
||||||
|
}
|
||||||
|
cleaned = append(cleaned, c)
|
||||||
|
case '.':
|
||||||
|
if i < len(b)-1 && b[i+1] == '_' {
|
||||||
|
return nil, newDecodeError(b[i+1:i+2], "cannot have underscore after decimal point")
|
||||||
|
}
|
||||||
|
if i > 0 && b[i-1] == '_' {
|
||||||
|
return nil, newDecodeError(b[i-1:i], "cannot have underscore before decimal point")
|
||||||
|
}
|
||||||
|
cleaned = append(cleaned, c)
|
||||||
|
default:
|
||||||
|
before = true
|
||||||
|
cleaned = append(cleaned, c)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return cleaned, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// isValidDate checks if a provided date is a date that exists.
|
||||||
|
func isValidDate(year int, month int, day int) bool {
|
||||||
|
return month > 0 && month < 13 && day > 0 && day <= daysIn(month, year)
|
||||||
|
}
|
||||||
|
|
||||||
|
// daysBefore[m] counts the number of days in a non-leap year
|
||||||
|
// before month m begins. There is an entry for m=12, counting
|
||||||
|
// the number of days before January of next year (365).
|
||||||
|
var daysBefore = [...]int32{
|
||||||
|
0,
|
||||||
|
31,
|
||||||
|
31 + 28,
|
||||||
|
31 + 28 + 31,
|
||||||
|
31 + 28 + 31 + 30,
|
||||||
|
31 + 28 + 31 + 30 + 31,
|
||||||
|
31 + 28 + 31 + 30 + 31 + 30,
|
||||||
|
31 + 28 + 31 + 30 + 31 + 30 + 31,
|
||||||
|
31 + 28 + 31 + 30 + 31 + 30 + 31 + 31,
|
||||||
|
31 + 28 + 31 + 30 + 31 + 30 + 31 + 31 + 30,
|
||||||
|
31 + 28 + 31 + 30 + 31 + 30 + 31 + 31 + 30 + 31,
|
||||||
|
31 + 28 + 31 + 30 + 31 + 30 + 31 + 31 + 30 + 31 + 30,
|
||||||
|
31 + 28 + 31 + 30 + 31 + 30 + 31 + 31 + 30 + 31 + 30 + 31,
|
||||||
|
}
|
||||||
|
|
||||||
|
func daysIn(m int, year int) int {
|
||||||
|
if m == 2 && isLeap(year) {
|
||||||
|
return 29
|
||||||
|
}
|
||||||
|
return int(daysBefore[m] - daysBefore[m-1])
|
||||||
|
}
|
||||||
|
|
||||||
|
func isLeap(year int) bool {
|
||||||
|
return year%4 == 0 && (year%100 != 0 || year%400 == 0)
|
||||||
|
}
|
||||||
|
|||||||
@@ -27,7 +27,7 @@ type DecodeError struct {
|
|||||||
// corresponding field in the target value. It contains all the missing fields
|
// corresponding field in the target value. It contains all the missing fields
|
||||||
// in Errors.
|
// in Errors.
|
||||||
//
|
//
|
||||||
// Emitted by Decoder when SetStrict(true) was called.
|
// Emitted by Decoder when DisallowUnknownFields() was called.
|
||||||
type StrictMissingError struct {
|
type StrictMissingError struct {
|
||||||
// One error per field that could not be found.
|
// One error per field that could not be found.
|
||||||
Errors []DecodeError
|
Errors []DecodeError
|
||||||
@@ -103,6 +103,7 @@ func (e *DecodeError) Key() Key {
|
|||||||
//
|
//
|
||||||
// The function copies all bytes used in DecodeError, so that document and
|
// The function copies all bytes used in DecodeError, so that document and
|
||||||
// highlight can be freely deallocated.
|
// highlight can be freely deallocated.
|
||||||
|
//
|
||||||
//nolint:funlen
|
//nolint:funlen
|
||||||
func wrapDecodeError(document []byte, de *decodeError) *DecodeError {
|
func wrapDecodeError(document []byte, de *decodeError) *DecodeError {
|
||||||
offset := danger.SubsliceOffset(document, de.highlight)
|
offset := danger.SubsliceOffset(document, de.highlight)
|
||||||
@@ -116,6 +117,7 @@ func wrapDecodeError(document []byte, de *decodeError) *DecodeError {
|
|||||||
maxLine := errLine + len(after) - 1
|
maxLine := errLine + len(after) - 1
|
||||||
lineColumnWidth := len(strconv.Itoa(maxLine))
|
lineColumnWidth := len(strconv.Itoa(maxLine))
|
||||||
|
|
||||||
|
// Write the lines of context strictly before the error.
|
||||||
for i := len(before) - 1; i > 0; i-- {
|
for i := len(before) - 1; i > 0; i-- {
|
||||||
line := errLine - i
|
line := errLine - i
|
||||||
buf.WriteString(formatLineNumber(line, lineColumnWidth))
|
buf.WriteString(formatLineNumber(line, lineColumnWidth))
|
||||||
@@ -129,6 +131,8 @@ func wrapDecodeError(document []byte, de *decodeError) *DecodeError {
|
|||||||
buf.WriteRune('\n')
|
buf.WriteRune('\n')
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Write the document line that contains the error.
|
||||||
|
|
||||||
buf.WriteString(formatLineNumber(errLine, lineColumnWidth))
|
buf.WriteString(formatLineNumber(errLine, lineColumnWidth))
|
||||||
buf.WriteString("| ")
|
buf.WriteString("| ")
|
||||||
|
|
||||||
@@ -143,6 +147,10 @@ func wrapDecodeError(document []byte, de *decodeError) *DecodeError {
|
|||||||
}
|
}
|
||||||
|
|
||||||
buf.WriteRune('\n')
|
buf.WriteRune('\n')
|
||||||
|
|
||||||
|
// Write the line with the error message itself (so it does not have a line
|
||||||
|
// number).
|
||||||
|
|
||||||
buf.WriteString(strings.Repeat(" ", lineColumnWidth))
|
buf.WriteString(strings.Repeat(" ", lineColumnWidth))
|
||||||
buf.WriteString("| ")
|
buf.WriteString("| ")
|
||||||
|
|
||||||
@@ -157,6 +165,8 @@ func wrapDecodeError(document []byte, de *decodeError) *DecodeError {
|
|||||||
buf.WriteString(errMessage)
|
buf.WriteString(errMessage)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Write the lines of context strictly after the error.
|
||||||
|
|
||||||
for i := 1; i < len(after); i++ {
|
for i := 1; i < len(after); i++ {
|
||||||
buf.WriteRune('\n')
|
buf.WriteRune('\n')
|
||||||
line := errLine + i
|
line := errLine + i
|
||||||
@@ -230,7 +240,7 @@ forward:
|
|||||||
rest = rest[o+1:]
|
rest = rest[o+1:]
|
||||||
o = 0
|
o = 0
|
||||||
|
|
||||||
case o == len(rest)-1 && o > 0:
|
case o == len(rest)-1:
|
||||||
// add last line only if it's non-empty
|
// add last line only if it's non-empty
|
||||||
afterLines = append(afterLines, rest)
|
afterLines = append(afterLines, rest)
|
||||||
|
|
||||||
|
|||||||
+13
-6
@@ -148,6 +148,13 @@ line 5`,
|
|||||||
6|
|
6|
|
||||||
7| line 4`,
|
7| line 4`,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
desc: "handle remainder of the error line when there is only one line",
|
||||||
|
doc: [3]string{`P=`, `[`, `#`},
|
||||||
|
msg: "array is incomplete",
|
||||||
|
expected: `1| P=[#
|
||||||
|
| ~ array is incomplete`,
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, e := range examples {
|
for _, e := range examples {
|
||||||
@@ -205,12 +212,12 @@ func ExampleDecodeError() {
|
|||||||
|
|
||||||
fmt.Println(err)
|
fmt.Println(err)
|
||||||
|
|
||||||
//nolint:errorlint
|
var derr *DecodeError
|
||||||
de := err.(*DecodeError)
|
if errors.As(err, &derr) {
|
||||||
fmt.Println(de.String())
|
fmt.Println(derr.String())
|
||||||
|
row, col := derr.Position()
|
||||||
row, col := de.Position()
|
fmt.Println("error occurred at row", row, "column", col)
|
||||||
fmt.Println("error occurred at row", row, "column", col)
|
}
|
||||||
// Output:
|
// Output:
|
||||||
// toml: number must have at least one digit between underscores
|
// toml: number must have at least one digit between underscores
|
||||||
// 1| name = 123__456
|
// 1| name = 123__456
|
||||||
|
|||||||
@@ -0,0 +1,56 @@
|
|||||||
|
//go:build go1.18 || go1.19
|
||||||
|
// +build go1.18 go1.19
|
||||||
|
|
||||||
|
package toml_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"io/ioutil"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
)
|
||||||
|
|
||||||
|
func FuzzUnmarshal(f *testing.F) {
|
||||||
|
file, err := ioutil.ReadFile("benchmark/benchmark.toml")
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
f.Add(file)
|
||||||
|
|
||||||
|
f.Fuzz(func(t *testing.T, b []byte) {
|
||||||
|
if strings.Contains(string(b), "nan") {
|
||||||
|
// Current limitation of testify.
|
||||||
|
// https://github.com/stretchr/testify/issues/624
|
||||||
|
t.Skip("can't compare NaNs")
|
||||||
|
}
|
||||||
|
|
||||||
|
t.Log("INITIAL DOCUMENT ===========================")
|
||||||
|
t.Log(string(b))
|
||||||
|
|
||||||
|
var v interface{}
|
||||||
|
err := toml.Unmarshal(b, &v)
|
||||||
|
if err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
t.Log("DECODED VALUE ===========================")
|
||||||
|
t.Logf("%#+v", v)
|
||||||
|
|
||||||
|
encoded, err := toml.Marshal(v)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("cannot marshal unmarshaled document: %s", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
t.Log("ENCODED DOCUMENT ===========================")
|
||||||
|
t.Log(string(encoded))
|
||||||
|
|
||||||
|
var v2 interface{}
|
||||||
|
err = toml.Unmarshal(encoded, &v2)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("failed round trip: %s", err)
|
||||||
|
}
|
||||||
|
require.Equal(t, v, v2)
|
||||||
|
})
|
||||||
|
}
|
||||||
@@ -1,6 +1,5 @@
|
|||||||
module github.com/pelletier/go-toml/v2
|
module github.com/pelletier/go-toml/v2
|
||||||
|
|
||||||
go 1.15
|
go 1.16
|
||||||
|
|
||||||
// latest (v1.7.0) doesn't have the fix for time.Time
|
require github.com/stretchr/testify v1.8.0
|
||||||
require github.com/stretchr/testify v1.7.1-0.20210427113832-6241f9ab9942
|
|
||||||
|
|||||||
@@ -1,11 +1,15 @@
|
|||||||
github.com/davecgh/go-spew v1.1.0 h1:ZDRjVQ15GmhC3fiQ8ni8+OwkZQO4DARzQgrnXU1Liz8=
|
|
||||||
github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
||||||
|
github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c=
|
||||||
|
github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
||||||
github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM=
|
github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM=
|
||||||
github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4=
|
github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4=
|
||||||
github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
||||||
github.com/stretchr/testify v1.7.1-0.20210427113832-6241f9ab9942 h1:t0lM6y/M5IiUZyvbBTcngso8SZEZICH7is9B6g/obVU=
|
github.com/stretchr/objx v0.4.0/go.mod h1:YvHI0jy2hoMjB+UWwv71VJQ9isScKT/TqJzVSSt89Yw=
|
||||||
github.com/stretchr/testify v1.7.1-0.20210427113832-6241f9ab9942/go.mod h1:6Fq8oRcR53rry900zMqJjRRixrwX3KX962/h/Wwjteg=
|
github.com/stretchr/testify v1.7.1/go.mod h1:6Fq8oRcR53rry900zMqJjRRixrwX3KX962/h/Wwjteg=
|
||||||
|
github.com/stretchr/testify v1.8.0 h1:pSgiaMZlXftHpm5L7V1+rVB+AZJydKsMxsQBIJw4PKk=
|
||||||
|
github.com/stretchr/testify v1.8.0/go.mod h1:yNjHg4UonilssWZ8iaSj1OCr/vHnekPRkoO+kdMU+MU=
|
||||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405 h1:yhCVgyC4o1eVCa2tZl7eS0r+SDo693bJlVdllGtEeKM=
|
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405 h1:yhCVgyC4o1eVCa2tZl7eS0r+SDo693bJlVdllGtEeKM=
|
||||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
||||||
gopkg.in/yaml.v3 v3.0.0-20200313102051-9f266ea9e77c h1:dUUwHk2QECo/6vqA44rthZ8ie2QXMNeKRTHCNY2nXvo=
|
|
||||||
gopkg.in/yaml.v3 v3.0.0-20200313102051-9f266ea9e77c/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
|
gopkg.in/yaml.v3 v3.0.0-20200313102051-9f266ea9e77c/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
|
||||||
|
gopkg.in/yaml.v3 v3.0.1 h1:fxVm/GzAzEWqLHuvctI91KS9hhNmmWOoWu0XTYJS7CA=
|
||||||
|
gopkg.in/yaml.v3 v3.0.1/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
|
||||||
|
|||||||
+28
-36
@@ -11,17 +11,17 @@ import (
|
|||||||
//
|
//
|
||||||
// For example:
|
// For example:
|
||||||
//
|
//
|
||||||
// it := n.Children()
|
// it := n.Children()
|
||||||
// for it.Next() {
|
// for it.Next() {
|
||||||
// it.Node()
|
// it.Node()
|
||||||
// }
|
// }
|
||||||
type Iterator struct {
|
type Iterator struct {
|
||||||
started bool
|
started bool
|
||||||
node *Node
|
node *Node
|
||||||
}
|
}
|
||||||
|
|
||||||
// Next moves the iterator forward and returns true if points to a node, false
|
// Next moves the iterator forward and returns true if points to a
|
||||||
// otherwise.
|
// node, false otherwise.
|
||||||
func (c *Iterator) Next() bool {
|
func (c *Iterator) Next() bool {
|
||||||
if !c.started {
|
if !c.started {
|
||||||
c.started = true
|
c.started = true
|
||||||
@@ -31,8 +31,8 @@ func (c *Iterator) Next() bool {
|
|||||||
return c.node.Valid()
|
return c.node.Valid()
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsLast returns true if the current node of the iterator is the last one.
|
// IsLast returns true if the current node of the iterator is the last
|
||||||
// Subsequent call to Next() will return false.
|
// one. Subsequent call to Next() will return false.
|
||||||
func (c *Iterator) IsLast() bool {
|
func (c *Iterator) IsLast() bool {
|
||||||
return c.node.next == 0
|
return c.node.next == 0
|
||||||
}
|
}
|
||||||
@@ -62,20 +62,20 @@ func (r *Root) at(idx Reference) *Node {
|
|||||||
return &r.nodes[idx]
|
return &r.nodes[idx]
|
||||||
}
|
}
|
||||||
|
|
||||||
// Arrays have one child per element in the array.
|
// Arrays have one child per element in the array. InlineTables have
|
||||||
// InlineTables have one child per key-value pair in the table.
|
// one child per key-value pair in the table. KeyValues have at least
|
||||||
// KeyValues have at least two children. The first one is the value. The
|
// two children. The first one is the value. The rest make a
|
||||||
// rest make a potentially dotted key.
|
// potentially dotted key. Table and Array table have one child per
|
||||||
// Table and Array table have one child per element of the key they
|
// element of the key they represent (same as KeyValue, but without
|
||||||
// represent (same as KeyValue, but without the last node being the value).
|
// the last node being the value).
|
||||||
// children []Node
|
|
||||||
type Node struct {
|
type Node struct {
|
||||||
Kind Kind
|
Kind Kind
|
||||||
Raw Range // Raw bytes from the input.
|
Raw Range // Raw bytes from the input.
|
||||||
Data []byte // Node value (could be either allocated or referencing the input).
|
Data []byte // Node value (either allocated or referencing the input).
|
||||||
|
|
||||||
// References to other nodes, as offsets in the backing array from this
|
// References to other nodes, as offsets in the backing array
|
||||||
// node. References can go backward, so those can be negative.
|
// from this node. References can go backward, so those can be
|
||||||
|
// negative.
|
||||||
next int // 0 if last element
|
next int // 0 if last element
|
||||||
child int // 0 if no child
|
child int // 0 if no child
|
||||||
}
|
}
|
||||||
@@ -85,8 +85,8 @@ type Range struct {
|
|||||||
Length uint32
|
Length uint32
|
||||||
}
|
}
|
||||||
|
|
||||||
// Next returns a copy of the next node, or an invalid Node if there is no
|
// Next returns a copy of the next node, or an invalid Node if there
|
||||||
// next node.
|
// is no next node.
|
||||||
func (n *Node) Next() *Node {
|
func (n *Node) Next() *Node {
|
||||||
if n.next == 0 {
|
if n.next == 0 {
|
||||||
return nil
|
return nil
|
||||||
@@ -96,9 +96,9 @@ func (n *Node) Next() *Node {
|
|||||||
return (*Node)(danger.Stride(ptr, size, n.next))
|
return (*Node)(danger.Stride(ptr, size, n.next))
|
||||||
}
|
}
|
||||||
|
|
||||||
// Child returns a copy of the first child node of this node. Other children
|
// Child returns a copy of the first child node of this node. Other
|
||||||
// can be accessed calling Next on the first child.
|
// children can be accessed calling Next on the first child. Returns
|
||||||
// Returns an invalid Node if there is none.
|
// an invalid Node if there is none.
|
||||||
func (n *Node) Child() *Node {
|
func (n *Node) Child() *Node {
|
||||||
if n.child == 0 {
|
if n.child == 0 {
|
||||||
return nil
|
return nil
|
||||||
@@ -113,10 +113,9 @@ func (n *Node) Valid() bool {
|
|||||||
return n != nil
|
return n != nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Key returns the child nodes making the Key on a supported node. Panics
|
// Key returns the child nodes making the Key on a supported
|
||||||
// otherwise.
|
// node. Panics otherwise. They are guaranteed to be all be of the
|
||||||
// They are guaranteed to be all be of the Kind Key. A simple key would return
|
// Kind Key. A simple key would return just one element.
|
||||||
// just one element.
|
|
||||||
func (n *Node) Key() Iterator {
|
func (n *Node) Key() Iterator {
|
||||||
switch n.Kind {
|
switch n.Kind {
|
||||||
case KeyValue:
|
case KeyValue:
|
||||||
@@ -133,10 +132,9 @@ func (n *Node) Key() Iterator {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Value returns a pointer to the value node of a KeyValue.
|
// Value returns a pointer to the value node of a KeyValue.
|
||||||
// Guaranteed to be non-nil.
|
// Guaranteed to be non-nil. Panics if not called on a KeyValue node,
|
||||||
// Panics if not called on a KeyValue node, or if the Children are malformed.
|
// or if the Children are malformed.
|
||||||
func (n *Node) Value() *Node {
|
func (n *Node) Value() *Node {
|
||||||
assertKind(KeyValue, *n)
|
|
||||||
return n.Child()
|
return n.Child()
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -144,9 +142,3 @@ func (n *Node) Value() *Node {
|
|||||||
func (n *Node) Children() Iterator {
|
func (n *Node) Children() Iterator {
|
||||||
return Iterator{node: n.Child()}
|
return Iterator{node: n.Child()}
|
||||||
}
|
}
|
||||||
|
|
||||||
func assertKind(k Kind, n Node) {
|
|
||||||
if n.Kind != k {
|
|
||||||
panic(fmt.Errorf("method was expecting a %s, not a %s", k, n.Kind))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -25,9 +25,9 @@ const (
|
|||||||
Float
|
Float
|
||||||
Integer
|
Integer
|
||||||
LocalDate
|
LocalDate
|
||||||
|
LocalTime
|
||||||
LocalDateTime
|
LocalDateTime
|
||||||
DateTime
|
DateTime
|
||||||
Time
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func (k Kind) String() string {
|
func (k Kind) String() string {
|
||||||
@@ -58,12 +58,12 @@ func (k Kind) String() string {
|
|||||||
return "Integer"
|
return "Integer"
|
||||||
case LocalDate:
|
case LocalDate:
|
||||||
return "LocalDate"
|
return "LocalDate"
|
||||||
|
case LocalTime:
|
||||||
|
return "LocalTime"
|
||||||
case LocalDateTime:
|
case LocalDateTime:
|
||||||
return "LocalDateTime"
|
return "LocalDateTime"
|
||||||
case DateTime:
|
case DateTime:
|
||||||
return "DateTime"
|
return "DateTime"
|
||||||
case Time:
|
|
||||||
return "Time"
|
|
||||||
}
|
}
|
||||||
panic(fmt.Errorf("Kind.String() not implemented for '%d'", k))
|
panic(fmt.Errorf("Kind.String() not implemented for '%d'", k))
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -0,0 +1,88 @@
|
|||||||
|
package cli
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"errors"
|
||||||
|
"flag"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"os"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
type ConvertFn func(r io.Reader, w io.Writer) error
|
||||||
|
|
||||||
|
type Program struct {
|
||||||
|
Usage string
|
||||||
|
Fn ConvertFn
|
||||||
|
// Inplace allows the command to take more than one file as argument and
|
||||||
|
// perform convertion in place on each provided file.
|
||||||
|
Inplace bool
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *Program) Execute() {
|
||||||
|
flag.Usage = func() { fmt.Fprintf(os.Stderr, p.Usage) }
|
||||||
|
flag.Parse()
|
||||||
|
os.Exit(p.main(flag.Args(), os.Stdin, os.Stdout, os.Stderr))
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *Program) main(files []string, input io.Reader, output, error io.Writer) int {
|
||||||
|
err := p.run(files, input, output)
|
||||||
|
if err != nil {
|
||||||
|
|
||||||
|
var derr *toml.DecodeError
|
||||||
|
if errors.As(err, &derr) {
|
||||||
|
fmt.Fprintln(error, derr.String())
|
||||||
|
row, col := derr.Position()
|
||||||
|
fmt.Fprintln(error, "error occurred at row", row, "column", col)
|
||||||
|
} else {
|
||||||
|
fmt.Fprintln(error, err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
return -1
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *Program) run(files []string, input io.Reader, output io.Writer) error {
|
||||||
|
if len(files) > 0 {
|
||||||
|
if p.Inplace {
|
||||||
|
return p.runAllFilesInPlace(files)
|
||||||
|
}
|
||||||
|
f, err := os.Open(files[0])
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
defer f.Close()
|
||||||
|
input = f
|
||||||
|
}
|
||||||
|
return p.Fn(input, output)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *Program) runAllFilesInPlace(files []string) error {
|
||||||
|
for _, path := range files {
|
||||||
|
err := p.runFileInPlace(path)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p *Program) runFileInPlace(path string) error {
|
||||||
|
in, err := ioutil.ReadFile(path)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
out := new(bytes.Buffer)
|
||||||
|
|
||||||
|
err = p.Fn(bytes.NewReader(in), out)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
return ioutil.WriteFile(path, out.Bytes(), 0600)
|
||||||
|
}
|
||||||
@@ -0,0 +1,172 @@
|
|||||||
|
package cli
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"os"
|
||||||
|
"path"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
)
|
||||||
|
|
||||||
|
func processMain(args []string, input io.Reader, stdout, stderr io.Writer, f ConvertFn) int {
|
||||||
|
p := Program{Fn: f}
|
||||||
|
return p.main(args, input, stdout, stderr)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProcessMainStdin(t *testing.T) {
|
||||||
|
stdout := new(bytes.Buffer)
|
||||||
|
stderr := new(bytes.Buffer)
|
||||||
|
input := strings.NewReader("this is the input")
|
||||||
|
|
||||||
|
exit := processMain([]string{}, input, stdout, stderr, func(r io.Reader, w io.Writer) error {
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
|
||||||
|
assert.Equal(t, 0, exit)
|
||||||
|
assert.Empty(t, stdout.String())
|
||||||
|
assert.Empty(t, stderr.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProcessMainStdinErr(t *testing.T) {
|
||||||
|
stdout := new(bytes.Buffer)
|
||||||
|
stderr := new(bytes.Buffer)
|
||||||
|
input := strings.NewReader("this is the input")
|
||||||
|
|
||||||
|
exit := processMain([]string{}, input, stdout, stderr, func(r io.Reader, w io.Writer) error {
|
||||||
|
return fmt.Errorf("something bad")
|
||||||
|
})
|
||||||
|
|
||||||
|
assert.Equal(t, -1, exit)
|
||||||
|
assert.Empty(t, stdout.String())
|
||||||
|
assert.NotEmpty(t, stderr.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProcessMainStdinDecodeErr(t *testing.T) {
|
||||||
|
stdout := new(bytes.Buffer)
|
||||||
|
stderr := new(bytes.Buffer)
|
||||||
|
input := strings.NewReader("this is the input")
|
||||||
|
|
||||||
|
exit := processMain([]string{}, input, stdout, stderr, func(r io.Reader, w io.Writer) error {
|
||||||
|
var v interface{}
|
||||||
|
return toml.Unmarshal([]byte(`qwe = 001`), &v)
|
||||||
|
})
|
||||||
|
|
||||||
|
assert.Equal(t, -1, exit)
|
||||||
|
assert.Empty(t, stdout.String())
|
||||||
|
assert.Contains(t, stderr.String(), "error occurred at")
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProcessMainFileExists(t *testing.T) {
|
||||||
|
tmpfile, err := ioutil.TempFile("", "example")
|
||||||
|
require.NoError(t, err)
|
||||||
|
defer os.Remove(tmpfile.Name())
|
||||||
|
_, err = tmpfile.Write([]byte(`some data`))
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
stdout := new(bytes.Buffer)
|
||||||
|
stderr := new(bytes.Buffer)
|
||||||
|
|
||||||
|
exit := processMain([]string{tmpfile.Name()}, nil, stdout, stderr, func(r io.Reader, w io.Writer) error {
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
|
||||||
|
assert.Equal(t, 0, exit)
|
||||||
|
assert.Empty(t, stdout.String())
|
||||||
|
assert.Empty(t, stderr.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProcessMainFileDoesNotExist(t *testing.T) {
|
||||||
|
stdout := new(bytes.Buffer)
|
||||||
|
stderr := new(bytes.Buffer)
|
||||||
|
|
||||||
|
exit := processMain([]string{"/lets/hope/this/does/not/exist"}, nil, stdout, stderr, func(r io.Reader, w io.Writer) error {
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
|
||||||
|
assert.Equal(t, -1, exit)
|
||||||
|
assert.Empty(t, stdout.String())
|
||||||
|
assert.NotEmpty(t, stderr.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProcessMainFilesInPlace(t *testing.T) {
|
||||||
|
dir, err := ioutil.TempDir("", "")
|
||||||
|
require.NoError(t, err)
|
||||||
|
defer os.RemoveAll(dir)
|
||||||
|
|
||||||
|
path1 := path.Join(dir, "file1")
|
||||||
|
path2 := path.Join(dir, "file2")
|
||||||
|
|
||||||
|
err = ioutil.WriteFile(path1, []byte("content 1"), 0600)
|
||||||
|
require.NoError(t, err)
|
||||||
|
err = ioutil.WriteFile(path2, []byte("content 2"), 0600)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
p := Program{
|
||||||
|
Fn: dummyFileFn,
|
||||||
|
Inplace: true,
|
||||||
|
}
|
||||||
|
|
||||||
|
exit := p.main([]string{path1, path2}, os.Stdin, os.Stdout, os.Stderr)
|
||||||
|
|
||||||
|
require.Equal(t, 0, exit)
|
||||||
|
|
||||||
|
v1, err := ioutil.ReadFile(path1)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, "1", string(v1))
|
||||||
|
|
||||||
|
v2, err := ioutil.ReadFile(path2)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, "2", string(v2))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProcessMainFilesInPlaceErrRead(t *testing.T) {
|
||||||
|
p := Program{
|
||||||
|
Fn: dummyFileFn,
|
||||||
|
Inplace: true,
|
||||||
|
}
|
||||||
|
|
||||||
|
exit := p.main([]string{"/this/path/is/invalid"}, os.Stdin, os.Stdout, os.Stderr)
|
||||||
|
|
||||||
|
require.Equal(t, -1, exit)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestProcessMainFilesInPlaceFailFn(t *testing.T) {
|
||||||
|
dir, err := ioutil.TempDir("", "")
|
||||||
|
require.NoError(t, err)
|
||||||
|
defer os.RemoveAll(dir)
|
||||||
|
|
||||||
|
path1 := path.Join(dir, "file1")
|
||||||
|
|
||||||
|
err = ioutil.WriteFile(path1, []byte("content 1"), 0600)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
p := Program{
|
||||||
|
Fn: func(io.Reader, io.Writer) error { return fmt.Errorf("oh no") },
|
||||||
|
Inplace: true,
|
||||||
|
}
|
||||||
|
|
||||||
|
exit := p.main([]string{path1}, os.Stdin, os.Stdout, os.Stderr)
|
||||||
|
|
||||||
|
require.Equal(t, -1, exit)
|
||||||
|
|
||||||
|
v1, err := ioutil.ReadFile(path1)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, "content 1", string(v1))
|
||||||
|
}
|
||||||
|
|
||||||
|
func dummyFileFn(r io.Reader, w io.Writer) error {
|
||||||
|
b, err := ioutil.ReadAll(r)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
v := strings.SplitN(string(b), " ", 2)[1]
|
||||||
|
_, err = w.Write([]byte(v))
|
||||||
|
return err
|
||||||
|
}
|
||||||
@@ -0,0 +1,23 @@
|
|||||||
|
package danger
|
||||||
|
|
||||||
|
import (
|
||||||
|
"reflect"
|
||||||
|
"unsafe"
|
||||||
|
)
|
||||||
|
|
||||||
|
// typeID is used as key in encoder and decoder caches to enable using
|
||||||
|
// the optimize runtime.mapaccess2_fast64 function instead of the more
|
||||||
|
// expensive lookup if we were to use reflect.Type as map key.
|
||||||
|
//
|
||||||
|
// typeID holds the pointer to the reflect.Type value, which is unique
|
||||||
|
// in the program.
|
||||||
|
//
|
||||||
|
// https://github.com/segmentio/encoding/blob/master/json/codec.go#L59-L61
|
||||||
|
type TypeID unsafe.Pointer
|
||||||
|
|
||||||
|
func MakeTypeID(t reflect.Type) TypeID {
|
||||||
|
// reflect.Type has the fields:
|
||||||
|
// typ unsafe.Pointer
|
||||||
|
// ptr unsafe.Pointer
|
||||||
|
return TypeID((*[2]unsafe.Pointer)(unsafe.Pointer(&t))[1])
|
||||||
|
}
|
||||||
@@ -67,6 +67,7 @@ func TestDocMarshal(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
marshalTestToml := `title = 'TOML Marshal Testing'
|
marshalTestToml := `title = 'TOML Marshal Testing'
|
||||||
|
|
||||||
[basic_lists]
|
[basic_lists]
|
||||||
floats = [12.3, 45.6, 78.9]
|
floats = [12.3, 45.6, 78.9]
|
||||||
bools = [true, false, true]
|
bools = [true, false, true]
|
||||||
@@ -89,7 +90,6 @@ name = 'Second'
|
|||||||
[subdoc.first]
|
[subdoc.first]
|
||||||
name = 'First'
|
name = 'First'
|
||||||
|
|
||||||
|
|
||||||
[basic]
|
[basic]
|
||||||
uint = 5001
|
uint = 5001
|
||||||
bool = true
|
bool = true
|
||||||
@@ -101,9 +101,9 @@ date = 1979-05-27T07:32:00Z
|
|||||||
|
|
||||||
[[subdoclist]]
|
[[subdoclist]]
|
||||||
name = 'List.First'
|
name = 'List.First'
|
||||||
|
|
||||||
[[subdoclist]]
|
[[subdoclist]]
|
||||||
name = 'List.Second'
|
name = 'List.Second'
|
||||||
|
|
||||||
`
|
`
|
||||||
|
|
||||||
result, err := toml.Marshal(docData)
|
result, err := toml.Marshal(docData)
|
||||||
@@ -117,14 +117,15 @@ func TestBasicMarshalQuotedKey(t *testing.T) {
|
|||||||
|
|
||||||
expected := `'Z.string-àéù' = 'Hello'
|
expected := `'Z.string-àéù' = 'Hello'
|
||||||
'Yfloat-𝟘' = 3.5
|
'Yfloat-𝟘' = 3.5
|
||||||
|
|
||||||
['Xsubdoc-àéù']
|
['Xsubdoc-àéù']
|
||||||
String2 = 'One'
|
String2 = 'One'
|
||||||
|
|
||||||
[['W.sublist-𝟘']]
|
[['W.sublist-𝟘']]
|
||||||
String2 = 'Two'
|
String2 = 'Two'
|
||||||
|
|
||||||
[['W.sublist-𝟘']]
|
[['W.sublist-𝟘']]
|
||||||
String2 = 'Three'
|
String2 = 'Three'
|
||||||
|
|
||||||
`
|
`
|
||||||
|
|
||||||
require.Equal(t, string(expected), string(result))
|
require.Equal(t, string(expected), string(result))
|
||||||
@@ -159,8 +160,8 @@ bool = false
|
|||||||
int = 0
|
int = 0
|
||||||
string = ''
|
string = ''
|
||||||
stringlist = []
|
stringlist = []
|
||||||
[map]
|
|
||||||
|
|
||||||
|
[map]
|
||||||
`
|
`
|
||||||
|
|
||||||
require.Equal(t, string(expected), string(result))
|
require.Equal(t, string(expected), string(result))
|
||||||
|
|||||||
@@ -151,6 +151,7 @@ type quotedKeyMarshalTestStruct struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var quotedKeyMarshalTestData = quotedKeyMarshalTestStruct{
|
var quotedKeyMarshalTestData = quotedKeyMarshalTestStruct{
|
||||||
String: "Hello",
|
String: "Hello",
|
||||||
@@ -160,6 +161,7 @@ var quotedKeyMarshalTestData = quotedKeyMarshalTestStruct{
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var quotedKeyMarshalTestToml = []byte(`"Yfloat-𝟘" = 3.5
|
var quotedKeyMarshalTestToml = []byte(`"Yfloat-𝟘" = 3.5
|
||||||
"Z.string-àéù" = "Hello"
|
"Z.string-àéù" = "Hello"
|
||||||
@@ -272,6 +274,7 @@ var docData = testDoc{
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var mapTestDoc = testMapDoc{
|
var mapTestDoc = testMapDoc{
|
||||||
Title: "TOML Marshal Testing",
|
Title: "TOML Marshal Testing",
|
||||||
@@ -457,35 +460,6 @@ func TestEmptytomlUnmarshal(t *testing.T) {
|
|||||||
assert.Equal(t, emptyTestData, result)
|
assert.Equal(t, emptyTestData, result)
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestEmptyUnmarshalOmit(t *testing.T) {
|
|
||||||
t.Skipf("Have not figured yet if omitempty is a good idea")
|
|
||||||
|
|
||||||
type emptyMarshalTestStruct2 struct {
|
|
||||||
Title string `toml:"title"`
|
|
||||||
Bool bool `toml:"bool,omitempty"`
|
|
||||||
Int int `toml:"int, omitempty"`
|
|
||||||
String string `toml:"string,omitempty "`
|
|
||||||
StringList []string `toml:"stringlist,omitempty"`
|
|
||||||
Ptr *basicMarshalTestStruct `toml:"ptr,omitempty"`
|
|
||||||
Map map[string]string `toml:"map,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
emptyTestData2 := emptyMarshalTestStruct2{
|
|
||||||
Title: "Placeholder",
|
|
||||||
Bool: false,
|
|
||||||
Int: 0,
|
|
||||||
String: "",
|
|
||||||
StringList: []string{},
|
|
||||||
Ptr: nil,
|
|
||||||
Map: map[string]string{},
|
|
||||||
}
|
|
||||||
|
|
||||||
result := emptyMarshalTestStruct2{}
|
|
||||||
err := toml.Unmarshal(emptyTestToml, &result)
|
|
||||||
require.NoError(t, err)
|
|
||||||
assert.Equal(t, emptyTestData2, result)
|
|
||||||
}
|
|
||||||
|
|
||||||
type pointerMarshalTestStruct struct {
|
type pointerMarshalTestStruct struct {
|
||||||
Str *string
|
Str *string
|
||||||
List *[]string
|
List *[]string
|
||||||
@@ -588,10 +562,12 @@ func (c customMarshaler) MarshalTOML() ([]byte, error) {
|
|||||||
var customMarshalerData = customMarshaler{FirstName: "Sally", LastName: "Fields"}
|
var customMarshalerData = customMarshaler{FirstName: "Sally", LastName: "Fields"}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var customMarshalerToml = []byte(`Sally Fields`)
|
var customMarshalerToml = []byte(`Sally Fields`)
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var nestedCustomMarshalerData = customMarshalerParent{
|
var nestedCustomMarshalerData = customMarshalerParent{
|
||||||
Self: customMarshaler{FirstName: "Maiku", LastName: "Suteda"},
|
Self: customMarshaler{FirstName: "Maiku", LastName: "Suteda"},
|
||||||
@@ -599,6 +575,7 @@ var nestedCustomMarshalerData = customMarshalerParent{
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var nestedCustomMarshalerToml = []byte(`friends = ["Sally Fields"]
|
var nestedCustomMarshalerToml = []byte(`friends = ["Sally Fields"]
|
||||||
me = "Maiku Suteda"
|
me = "Maiku Suteda"
|
||||||
@@ -640,6 +617,7 @@ func TestUnmarshalTextMarshaler(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once type and methods are used by a test
|
// TODO: Remove nolint once type and methods are used by a test
|
||||||
|
//
|
||||||
//nolint:unused
|
//nolint:unused
|
||||||
type precedentMarshaler struct {
|
type precedentMarshaler struct {
|
||||||
FirstName string
|
FirstName string
|
||||||
@@ -658,6 +636,7 @@ func (m precedentMarshaler) MarshalTOML() ([]byte, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once type and method are used by a test
|
// TODO: Remove nolint once type and method are used by a test
|
||||||
|
//
|
||||||
//nolint:unused
|
//nolint:unused
|
||||||
type customPointerMarshaler struct {
|
type customPointerMarshaler struct {
|
||||||
FirstName string
|
FirstName string
|
||||||
@@ -670,6 +649,7 @@ func (m *customPointerMarshaler) MarshalTOML() ([]byte, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once type and method are used by a test
|
// TODO: Remove nolint once type and method are used by a test
|
||||||
|
//
|
||||||
//nolint:unused
|
//nolint:unused
|
||||||
type textPointerMarshaler struct {
|
type textPointerMarshaler struct {
|
||||||
FirstName string
|
FirstName string
|
||||||
@@ -682,6 +662,7 @@ func (m *textPointerMarshaler) MarshalText() ([]byte, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var commentTestToml = []byte(`
|
var commentTestToml = []byte(`
|
||||||
# it's a comment on type
|
# it's a comment on type
|
||||||
@@ -719,6 +700,7 @@ type mapsTestStruct struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var mapsTestData = mapsTestStruct{
|
var mapsTestData = mapsTestStruct{
|
||||||
Simple: map[string]string{
|
Simple: map[string]string{
|
||||||
@@ -742,6 +724,7 @@ var mapsTestData = mapsTestStruct{
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var mapsTestToml = []byte(`
|
var mapsTestToml = []byte(`
|
||||||
[Other]
|
[Other]
|
||||||
@@ -764,6 +747,7 @@ var mapsTestToml = []byte(`
|
|||||||
`)
|
`)
|
||||||
|
|
||||||
// TODO: Remove nolint once type is used by a test
|
// TODO: Remove nolint once type is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused
|
//nolint:deadcode,unused
|
||||||
type structArrayNoTag struct {
|
type structArrayNoTag struct {
|
||||||
A struct {
|
A struct {
|
||||||
@@ -773,6 +757,7 @@ type structArrayNoTag struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var customTagTestToml = []byte(`
|
var customTagTestToml = []byte(`
|
||||||
[postgres]
|
[postgres]
|
||||||
@@ -787,6 +772,7 @@ var customTagTestToml = []byte(`
|
|||||||
`)
|
`)
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var customCommentTagTestToml = []byte(`
|
var customCommentTagTestToml = []byte(`
|
||||||
# db connection
|
# db connection
|
||||||
@@ -800,6 +786,7 @@ var customCommentTagTestToml = []byte(`
|
|||||||
`)
|
`)
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var customCommentedTagTestToml = []byte(`
|
var customCommentedTagTestToml = []byte(`
|
||||||
[postgres]
|
[postgres]
|
||||||
@@ -854,6 +841,7 @@ func TestUnmarshalTabInStringAndQuotedKey(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var customMultilineTagTestToml = []byte(`int_slice = [
|
var customMultilineTagTestToml = []byte(`int_slice = [
|
||||||
1,
|
1,
|
||||||
@@ -863,6 +851,7 @@ var customMultilineTagTestToml = []byte(`int_slice = [
|
|||||||
`)
|
`)
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var testDocBasicToml = []byte(`
|
var testDocBasicToml = []byte(`
|
||||||
[document]
|
[document]
|
||||||
@@ -875,12 +864,14 @@ var testDocBasicToml = []byte(`
|
|||||||
`)
|
`)
|
||||||
|
|
||||||
// TODO: Remove nolint once type is used by a test
|
// TODO: Remove nolint once type is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode
|
//nolint:deadcode
|
||||||
type testDocCustomTag struct {
|
type testDocCustomTag struct {
|
||||||
Doc testDocBasicsCustomTag `file:"document"`
|
Doc testDocBasicsCustomTag `file:"document"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once type is used by a test
|
// TODO: Remove nolint once type is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode
|
//nolint:deadcode
|
||||||
type testDocBasicsCustomTag struct {
|
type testDocBasicsCustomTag struct {
|
||||||
Bool bool `file:"bool_val"`
|
Bool bool `file:"bool_val"`
|
||||||
@@ -893,6 +884,7 @@ type testDocBasicsCustomTag struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,varcheck
|
//nolint:deadcode,varcheck
|
||||||
var testDocCustomTagData = testDocCustomTag{
|
var testDocCustomTagData = testDocCustomTag{
|
||||||
Doc: testDocBasicsCustomTag{
|
Doc: testDocBasicsCustomTag{
|
||||||
@@ -956,6 +948,29 @@ func TestUnmarshalMapWithTypedKey(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestUnmarshalTypeTableHeader(t *testing.T) {
|
||||||
|
testToml := []byte(`
|
||||||
|
[test]
|
||||||
|
a = 1
|
||||||
|
`)
|
||||||
|
|
||||||
|
type header string
|
||||||
|
var result map[header]map[string]int
|
||||||
|
err := toml.Unmarshal(testToml, &result)
|
||||||
|
if err != nil {
|
||||||
|
t.Errorf("Received unexpected error: %s", err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
expected := map[header]map[string]int{
|
||||||
|
"test": map[string]int{"a": 1},
|
||||||
|
}
|
||||||
|
|
||||||
|
if !reflect.DeepEqual(result, expected) {
|
||||||
|
t.Errorf("Bad unmarshal: expected %v, got %v", expected, result)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func TestUnmarshalNonPointer(t *testing.T) {
|
func TestUnmarshalNonPointer(t *testing.T) {
|
||||||
a := 1
|
a := 1
|
||||||
err := toml.Unmarshal([]byte{}, a)
|
err := toml.Unmarshal([]byte{}, a)
|
||||||
@@ -972,6 +987,7 @@ func TestUnmarshalInvalidPointerKind(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused
|
//nolint:deadcode,unused
|
||||||
type testDuration struct {
|
type testDuration struct {
|
||||||
Nanosec time.Duration `toml:"nanosec"`
|
Nanosec time.Duration `toml:"nanosec"`
|
||||||
@@ -986,6 +1002,7 @@ type testDuration struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var testDurationToml = []byte(`
|
var testDurationToml = []byte(`
|
||||||
nanosec = "1ns"
|
nanosec = "1ns"
|
||||||
@@ -1000,6 +1017,7 @@ a_string = "15s"
|
|||||||
`)
|
`)
|
||||||
|
|
||||||
// TODO: Remove nolint once var is used by a test
|
// TODO: Remove nolint once var is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused,varcheck
|
//nolint:deadcode,unused,varcheck
|
||||||
var testDurationToml2 = []byte(`a_string = "15s"
|
var testDurationToml2 = []byte(`a_string = "15s"
|
||||||
hour = "1h0m0s"
|
hour = "1h0m0s"
|
||||||
@@ -1013,6 +1031,7 @@ sec = "1s"
|
|||||||
`)
|
`)
|
||||||
|
|
||||||
// TODO: Remove nolint once type is used by a test
|
// TODO: Remove nolint once type is used by a test
|
||||||
|
//
|
||||||
//nolint:deadcode,unused
|
//nolint:deadcode,unused
|
||||||
type testBadDuration struct {
|
type testBadDuration struct {
|
||||||
Val time.Duration `toml:"val"`
|
Val time.Duration `toml:"val"`
|
||||||
@@ -1487,12 +1506,12 @@ func TestUnmarshalLocalDateTime(t *testing.T) {
|
|||||||
name: "normal",
|
name: "normal",
|
||||||
in: "1979-05-27T07:32:00",
|
in: "1979-05-27T07:32:00",
|
||||||
out: toml.LocalDateTime{
|
out: toml.LocalDateTime{
|
||||||
Date: toml.LocalDate{
|
LocalDate: toml.LocalDate{
|
||||||
Year: 1979,
|
Year: 1979,
|
||||||
Month: 5,
|
Month: 5,
|
||||||
Day: 27,
|
Day: 27,
|
||||||
},
|
},
|
||||||
Time: toml.LocalTime{
|
LocalTime: toml.LocalTime{
|
||||||
Hour: 7,
|
Hour: 7,
|
||||||
Minute: 32,
|
Minute: 32,
|
||||||
Second: 0,
|
Second: 0,
|
||||||
@@ -1504,16 +1523,17 @@ func TestUnmarshalLocalDateTime(t *testing.T) {
|
|||||||
name: "with nanoseconds",
|
name: "with nanoseconds",
|
||||||
in: "1979-05-27T00:32:00.999999",
|
in: "1979-05-27T00:32:00.999999",
|
||||||
out: toml.LocalDateTime{
|
out: toml.LocalDateTime{
|
||||||
Date: toml.LocalDate{
|
LocalDate: toml.LocalDate{
|
||||||
Year: 1979,
|
Year: 1979,
|
||||||
Month: 5,
|
Month: 5,
|
||||||
Day: 27,
|
Day: 27,
|
||||||
},
|
},
|
||||||
Time: toml.LocalTime{
|
LocalTime: toml.LocalTime{
|
||||||
Hour: 0,
|
Hour: 0,
|
||||||
Minute: 32,
|
Minute: 32,
|
||||||
Second: 0,
|
Second: 0,
|
||||||
Nanosecond: 999999000,
|
Nanosecond: 999999000,
|
||||||
|
Precision: 6,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
@@ -1551,26 +1571,26 @@ func TestUnmarshalLocalDateTime(t *testing.T) {
|
|||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if obj.Date.Year() != example.out.Date.Year {
|
if obj.Date.Year() != example.out.Year {
|
||||||
t.Errorf("expected year %d, got %d", example.out.Date.Year, obj.Date.Year())
|
t.Errorf("expected year %d, got %d", example.out.Year, obj.Date.Year())
|
||||||
}
|
}
|
||||||
if obj.Date.Month() != example.out.Date.Month {
|
if obj.Date.Month() != time.Month(example.out.Month) {
|
||||||
t.Errorf("expected month %d, got %d", example.out.Date.Month, obj.Date.Month())
|
t.Errorf("expected month %d, got %d", example.out.Month, obj.Date.Month())
|
||||||
}
|
}
|
||||||
if obj.Date.Day() != example.out.Date.Day {
|
if obj.Date.Day() != example.out.Day {
|
||||||
t.Errorf("expected day %d, got %d", example.out.Date.Day, obj.Date.Day())
|
t.Errorf("expected day %d, got %d", example.out.Day, obj.Date.Day())
|
||||||
}
|
}
|
||||||
if obj.Date.Hour() != example.out.Time.Hour {
|
if obj.Date.Hour() != example.out.Hour {
|
||||||
t.Errorf("expected hour %d, got %d", example.out.Time.Hour, obj.Date.Hour())
|
t.Errorf("expected hour %d, got %d", example.out.Hour, obj.Date.Hour())
|
||||||
}
|
}
|
||||||
if obj.Date.Minute() != example.out.Time.Minute {
|
if obj.Date.Minute() != example.out.Minute {
|
||||||
t.Errorf("expected minute %d, got %d", example.out.Time.Minute, obj.Date.Minute())
|
t.Errorf("expected minute %d, got %d", example.out.Minute, obj.Date.Minute())
|
||||||
}
|
}
|
||||||
if obj.Date.Second() != example.out.Time.Second {
|
if obj.Date.Second() != example.out.Second {
|
||||||
t.Errorf("expected second %d, got %d", example.out.Time.Second, obj.Date.Second())
|
t.Errorf("expected second %d, got %d", example.out.Second, obj.Date.Second())
|
||||||
}
|
}
|
||||||
if obj.Date.Nanosecond() != example.out.Time.Nanosecond {
|
if obj.Date.Nanosecond() != example.out.Nanosecond {
|
||||||
t.Errorf("expected nanoseconds %d, got %d", example.out.Time.Nanosecond, obj.Date.Nanosecond())
|
t.Errorf("expected nanoseconds %d, got %d", example.out.Nanosecond, obj.Date.Nanosecond())
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@@ -1600,6 +1620,7 @@ func TestUnmarshalLocalTime(t *testing.T) {
|
|||||||
Minute: 32,
|
Minute: 32,
|
||||||
Second: 0,
|
Second: 0,
|
||||||
Nanosecond: 999999000,
|
Nanosecond: 999999000,
|
||||||
|
Precision: 6,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
@@ -1952,7 +1973,7 @@ func decoder(doc string) *toml.Decoder {
|
|||||||
|
|
||||||
func strictDecoder(doc string) *toml.Decoder {
|
func strictDecoder(doc string) *toml.Decoder {
|
||||||
d := decoder(doc)
|
d := decoder(doc)
|
||||||
d.SetStrict(true)
|
d.DisallowUnknownFields()
|
||||||
return d
|
return d
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -0,0 +1,74 @@
|
|||||||
|
package testsuite
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"math"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
// addTag adds JSON tags to a data structure as expected by toml-test.
|
||||||
|
func addTag(key string, tomlData interface{}) interface{} {
|
||||||
|
// Switch on the data type.
|
||||||
|
switch orig := tomlData.(type) {
|
||||||
|
default:
|
||||||
|
//return map[string]interface{}{}
|
||||||
|
panic(fmt.Sprintf("Unknown type: %T", tomlData))
|
||||||
|
|
||||||
|
// A table: we don't need to add any tags, just recurse for every table
|
||||||
|
// entry.
|
||||||
|
case map[string]interface{}:
|
||||||
|
typed := make(map[string]interface{}, len(orig))
|
||||||
|
for k, v := range orig {
|
||||||
|
typed[k] = addTag(k, v)
|
||||||
|
}
|
||||||
|
return typed
|
||||||
|
|
||||||
|
// An array: we don't need to add any tags, just recurse for every table
|
||||||
|
// entry.
|
||||||
|
case []map[string]interface{}:
|
||||||
|
typed := make([]map[string]interface{}, len(orig))
|
||||||
|
for i, v := range orig {
|
||||||
|
typed[i] = addTag("", v).(map[string]interface{})
|
||||||
|
}
|
||||||
|
return typed
|
||||||
|
case []interface{}:
|
||||||
|
typed := make([]interface{}, len(orig))
|
||||||
|
for i, v := range orig {
|
||||||
|
typed[i] = addTag("", v)
|
||||||
|
}
|
||||||
|
return typed
|
||||||
|
|
||||||
|
// Datetime: tag as datetime.
|
||||||
|
case toml.LocalTime:
|
||||||
|
return tag("time-local", orig.String())
|
||||||
|
case toml.LocalDate:
|
||||||
|
return tag("date-local", orig.String())
|
||||||
|
case toml.LocalDateTime:
|
||||||
|
return tag("datetime-local", orig.String())
|
||||||
|
case time.Time:
|
||||||
|
return tag("datetime", orig.Format("2006-01-02T15:04:05.999999999Z07:00"))
|
||||||
|
|
||||||
|
// Tag primitive values: bool, string, int, and float64.
|
||||||
|
case bool:
|
||||||
|
return tag("bool", fmt.Sprintf("%v", orig))
|
||||||
|
case string:
|
||||||
|
return tag("string", orig)
|
||||||
|
case int64:
|
||||||
|
return tag("integer", fmt.Sprintf("%d", orig))
|
||||||
|
case float64:
|
||||||
|
// Special case for nan since NaN == NaN is false.
|
||||||
|
if math.IsNaN(orig) {
|
||||||
|
return tag("float", "nan")
|
||||||
|
}
|
||||||
|
return tag("float", fmt.Sprintf("%v", orig))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func tag(typeName string, data interface{}) map[string]interface{} {
|
||||||
|
return map[string]interface{}{
|
||||||
|
"type": typeName,
|
||||||
|
"value": data,
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -0,0 +1,244 @@
|
|||||||
|
package testsuite
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
func CmpJSON(t *testing.T, key string, want, have interface{}) {
|
||||||
|
switch w := want.(type) {
|
||||||
|
case map[string]interface{}:
|
||||||
|
cmpJSONMaps(t, key, w, have)
|
||||||
|
case []interface{}:
|
||||||
|
cmpJSONArrays(t, key, w, have)
|
||||||
|
default:
|
||||||
|
t.Errorf(
|
||||||
|
"Key '%s' in expected output should be a map or a list of maps, but it's a %T",
|
||||||
|
key, want)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func cmpJSONMaps(t *testing.T, key string, want map[string]interface{}, have interface{}) {
|
||||||
|
haveMap, ok := have.(map[string]interface{})
|
||||||
|
if !ok {
|
||||||
|
mismatch(t, key, "table", want, haveMap)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check to make sure both or neither are values.
|
||||||
|
if isValue(want) && !isValue(haveMap) {
|
||||||
|
t.Fatalf("Key '%s' is supposed to be a value, but the parser reports it as a table", key)
|
||||||
|
}
|
||||||
|
if !isValue(want) && isValue(haveMap) {
|
||||||
|
t.Fatalf("Key '%s' is supposed to be a table, but the parser reports it as a value", key)
|
||||||
|
}
|
||||||
|
if isValue(want) && isValue(haveMap) {
|
||||||
|
cmpJSONValues(t, key, want, haveMap)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check that the keys of each map are equivalent.
|
||||||
|
for k := range want {
|
||||||
|
if _, ok := haveMap[k]; !ok {
|
||||||
|
bunk := kjoin(key, k)
|
||||||
|
t.Fatalf("Could not find key '%s' in parser output.", bunk)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for k := range haveMap {
|
||||||
|
if _, ok := want[k]; !ok {
|
||||||
|
bunk := kjoin(key, k)
|
||||||
|
t.Fatalf("Could not find key '%s' in expected output.", bunk)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Okay, now make sure that each value is equivalent.
|
||||||
|
for k := range want {
|
||||||
|
CmpJSON(t, kjoin(key, k), want[k], haveMap[k])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func cmpJSONArrays(t *testing.T, key string, want, have interface{}) {
|
||||||
|
wantSlice, ok := want.([]interface{})
|
||||||
|
if !ok {
|
||||||
|
panic(fmt.Sprintf("'value' should be a JSON array when 'type=array', but it is a %T", want))
|
||||||
|
}
|
||||||
|
|
||||||
|
haveSlice, ok := have.([]interface{})
|
||||||
|
if !ok {
|
||||||
|
t.Fatalf("Malformed output from your encoder: 'value' is not a JSON array: %T", have)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(wantSlice) != len(haveSlice) {
|
||||||
|
t.Fatalf("Array lengths differ for key '%s':\n"+
|
||||||
|
" Expected: %d\n"+
|
||||||
|
" Your encoder: %d",
|
||||||
|
key, len(wantSlice), len(haveSlice))
|
||||||
|
}
|
||||||
|
for i := 0; i < len(wantSlice); i++ {
|
||||||
|
CmpJSON(t, key, wantSlice[i], haveSlice[i])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func cmpJSONValues(t *testing.T, key string, want, have map[string]interface{}) {
|
||||||
|
wantType, ok := want["type"].(string)
|
||||||
|
if !ok {
|
||||||
|
panic(fmt.Sprintf("'type' should be a string, but it is a %T", want["type"]))
|
||||||
|
}
|
||||||
|
|
||||||
|
haveType, ok := have["type"].(string)
|
||||||
|
if !ok {
|
||||||
|
t.Fatalf("Malformed output from your encoder: 'type' is not a string: %T", have["type"])
|
||||||
|
}
|
||||||
|
|
||||||
|
if wantType != haveType {
|
||||||
|
valMismatch(t, key, wantType, haveType, want, have)
|
||||||
|
}
|
||||||
|
|
||||||
|
// If this is an array, then we've got to do some work to check equality.
|
||||||
|
if wantType == "array" {
|
||||||
|
cmpJSONArrays(t, key, want, have)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Atomic values are always strings
|
||||||
|
wantVal, ok := want["value"].(string)
|
||||||
|
if !ok {
|
||||||
|
panic(fmt.Sprintf("'value' %v should be a string, but it is a %[1]T", want["value"]))
|
||||||
|
}
|
||||||
|
|
||||||
|
haveVal, ok := have["value"].(string)
|
||||||
|
if !ok {
|
||||||
|
panic(fmt.Sprintf("Malformed output from your encoder: %T is not a string", have["value"]))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Excepting floats and datetimes, other values can be compared as strings.
|
||||||
|
switch wantType {
|
||||||
|
case "float":
|
||||||
|
cmpFloats(t, key, wantVal, haveVal)
|
||||||
|
case "datetime", "datetime-local", "date-local", "time-local":
|
||||||
|
cmpAsDatetimes(t, key, wantType, wantVal, haveVal)
|
||||||
|
default:
|
||||||
|
cmpAsStrings(t, key, wantVal, haveVal)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func cmpAsStrings(t *testing.T, key string, want, have string) {
|
||||||
|
if want != have {
|
||||||
|
t.Fatalf("Values for key '%s' don't match:\n"+
|
||||||
|
" Expected: %s\n"+
|
||||||
|
" Your encoder: %s",
|
||||||
|
key, want, have)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func cmpFloats(t *testing.T, key string, want, have string) {
|
||||||
|
// Special case for NaN, since NaN != NaN.
|
||||||
|
if strings.HasSuffix(want, "nan") || strings.HasSuffix(have, "nan") {
|
||||||
|
if want != have {
|
||||||
|
t.Fatalf("Values for key '%s' don't match:\n"+
|
||||||
|
" Expected: %v\n"+
|
||||||
|
" Your encoder: %v",
|
||||||
|
key, want, have)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
wantF, err := strconv.ParseFloat(want, 64)
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Sprintf("Could not read '%s' as a float value for key '%s'", want, key))
|
||||||
|
}
|
||||||
|
|
||||||
|
haveF, err := strconv.ParseFloat(have, 64)
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Sprintf("Malformed output from your encoder: key '%s' is not a float: '%s'", key, have))
|
||||||
|
}
|
||||||
|
|
||||||
|
if wantF != haveF {
|
||||||
|
t.Fatalf("Values for key '%s' don't match:\n"+
|
||||||
|
" Expected: %v\n"+
|
||||||
|
" Your encoder: %v",
|
||||||
|
key, wantF, haveF)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
var datetimeRepl = strings.NewReplacer(
|
||||||
|
" ", "T",
|
||||||
|
"t", "T",
|
||||||
|
"z", "Z")
|
||||||
|
|
||||||
|
var layouts = map[string]string{
|
||||||
|
"datetime": time.RFC3339Nano,
|
||||||
|
"datetime-local": "2006-01-02T15:04:05.999999999",
|
||||||
|
"date-local": "2006-01-02",
|
||||||
|
"time-local": "15:04:05",
|
||||||
|
}
|
||||||
|
|
||||||
|
func cmpAsDatetimes(t *testing.T, key string, kind, want, have string) {
|
||||||
|
layout, ok := layouts[kind]
|
||||||
|
if !ok {
|
||||||
|
panic("should never happen")
|
||||||
|
}
|
||||||
|
|
||||||
|
wantT, err := time.Parse(layout, datetimeRepl.Replace(want))
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Sprintf("Could not read '%s' as a datetime value for key '%s'", want, key))
|
||||||
|
}
|
||||||
|
|
||||||
|
haveT, err := time.Parse(layout, datetimeRepl.Replace(want))
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("Malformed output from your encoder: key '%s' is not a datetime: '%s'", key, have)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if !wantT.Equal(haveT) {
|
||||||
|
t.Fatalf("Values for key '%s' don't match:\n"+
|
||||||
|
" Expected: %v\n"+
|
||||||
|
" Your encoder: %v",
|
||||||
|
key, wantT, haveT)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func cmpAsDatetimesLocal(t *testing.T, key string, want, have string) {
|
||||||
|
if datetimeRepl.Replace(want) != datetimeRepl.Replace(have) {
|
||||||
|
t.Fatalf("Values for key '%s' don't match:\n"+
|
||||||
|
" Expected: %v\n"+
|
||||||
|
" Your encoder: %v",
|
||||||
|
key, want, have)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func kjoin(old, key string) string {
|
||||||
|
if len(old) == 0 {
|
||||||
|
return key
|
||||||
|
}
|
||||||
|
return old + "." + key
|
||||||
|
}
|
||||||
|
|
||||||
|
func isValue(m map[string]interface{}) bool {
|
||||||
|
if len(m) != 2 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
if _, ok := m["type"]; !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
if _, ok := m["value"]; !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
func mismatch(t *testing.T, key string, wantType string, want, have interface{}) {
|
||||||
|
t.Fatalf("Key '%s' is not an %s but %[4]T:\n"+
|
||||||
|
" Expected: %#[3]v\n"+
|
||||||
|
" Your encoder: %#[4]v",
|
||||||
|
key, wantType, want, have)
|
||||||
|
}
|
||||||
|
|
||||||
|
func valMismatch(t *testing.T, key string, wantType, haveType string, want, have interface{}) {
|
||||||
|
t.Fatalf("Key '%s' is not an %s but %s:\n"+
|
||||||
|
" Expected: %#[3]v\n"+
|
||||||
|
" Your encoder: %#[4]v",
|
||||||
|
key, wantType, want, have)
|
||||||
|
}
|
||||||
@@ -0,0 +1,69 @@
|
|||||||
|
package testsuite
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
type parser struct{}
|
||||||
|
|
||||||
|
func (p parser) Decode(input string) (output string, outputIsError bool, retErr error) {
|
||||||
|
defer func() {
|
||||||
|
if r := recover(); r != nil {
|
||||||
|
switch rr := r.(type) {
|
||||||
|
case error:
|
||||||
|
retErr = rr
|
||||||
|
default:
|
||||||
|
retErr = fmt.Errorf("%s", rr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
var v interface{}
|
||||||
|
|
||||||
|
if err := toml.Unmarshal([]byte(input), &v); err != nil {
|
||||||
|
return err.Error(), true, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
j, err := json.MarshalIndent(addTag("", v), "", " ")
|
||||||
|
if err != nil {
|
||||||
|
return "", false, retErr
|
||||||
|
}
|
||||||
|
|
||||||
|
return string(j), false, retErr
|
||||||
|
}
|
||||||
|
|
||||||
|
func (p parser) Encode(input string) (output string, outputIsError bool, retErr error) {
|
||||||
|
defer func() {
|
||||||
|
if r := recover(); r != nil {
|
||||||
|
switch rr := r.(type) {
|
||||||
|
case error:
|
||||||
|
retErr = rr
|
||||||
|
default:
|
||||||
|
retErr = fmt.Errorf("%s", rr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
var tmp interface{}
|
||||||
|
err := json.Unmarshal([]byte(input), &tmp)
|
||||||
|
if err != nil {
|
||||||
|
return "", false, err
|
||||||
|
}
|
||||||
|
|
||||||
|
rm, err := rmTag(tmp)
|
||||||
|
if err != nil {
|
||||||
|
return err.Error(), true, retErr
|
||||||
|
}
|
||||||
|
|
||||||
|
buf := new(bytes.Buffer)
|
||||||
|
err = toml.NewEncoder(buf).Encode(rm)
|
||||||
|
if err != nil {
|
||||||
|
return err.Error(), true, retErr
|
||||||
|
}
|
||||||
|
|
||||||
|
return buf.String(), false, retErr
|
||||||
|
}
|
||||||
@@ -0,0 +1,110 @@
|
|||||||
|
package testsuite
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"strconv"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Remove JSON tags to a data structure as returned by toml-test.
|
||||||
|
func rmTag(typedJson interface{}) (interface{}, error) {
|
||||||
|
// Check if key is in the table m.
|
||||||
|
in := func(key string, m map[string]interface{}) bool {
|
||||||
|
_, ok := m[key]
|
||||||
|
return ok
|
||||||
|
}
|
||||||
|
|
||||||
|
// Switch on the data type.
|
||||||
|
switch v := typedJson.(type) {
|
||||||
|
|
||||||
|
// Object: this can either be a TOML table or a primitive with tags.
|
||||||
|
case map[string]interface{}:
|
||||||
|
// This value represents a primitive: remove the tags and return just
|
||||||
|
// the primitive value.
|
||||||
|
if len(v) == 2 && in("type", v) && in("value", v) {
|
||||||
|
ut, err := untag(v)
|
||||||
|
if err != nil {
|
||||||
|
return ut, fmt.Errorf("tag.Remove: %w", err)
|
||||||
|
}
|
||||||
|
return ut, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Table: remove tags on all children.
|
||||||
|
m := make(map[string]interface{}, len(v))
|
||||||
|
for k, v2 := range v {
|
||||||
|
var err error
|
||||||
|
m[k], err = rmTag(v2)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return m, nil
|
||||||
|
|
||||||
|
// Array: remove tags from all itenm.
|
||||||
|
case []interface{}:
|
||||||
|
a := make([]interface{}, len(v))
|
||||||
|
for i := range v {
|
||||||
|
var err error
|
||||||
|
a[i], err = rmTag(v[i])
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return a, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// The top level must be an object or array.
|
||||||
|
return nil, fmt.Errorf("unrecognized JSON format '%T'", typedJson)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return a primitive: read the "type" and convert the "value" to that.
|
||||||
|
func untag(typed map[string]interface{}) (interface{}, error) {
|
||||||
|
t := typed["type"].(string)
|
||||||
|
v := typed["value"].(string)
|
||||||
|
switch t {
|
||||||
|
case "string":
|
||||||
|
return v, nil
|
||||||
|
case "integer":
|
||||||
|
n, err := strconv.ParseInt(v, 10, 64)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("untag: %w", err)
|
||||||
|
}
|
||||||
|
return n, nil
|
||||||
|
case "float":
|
||||||
|
f, err := strconv.ParseFloat(v, 64)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("untag: %w", err)
|
||||||
|
}
|
||||||
|
return f, nil
|
||||||
|
case "datetime":
|
||||||
|
return parseTime(v, "2006-01-02T15:04:05.999999999Z07:00", false)
|
||||||
|
case "datetime-local":
|
||||||
|
return parseTime(v, "2006-01-02T15:04:05.999999999", true)
|
||||||
|
case "date-local":
|
||||||
|
return parseTime(v, "2006-01-02", true)
|
||||||
|
case "time-local":
|
||||||
|
return parseTime(v, "15:04:05.999999999", true)
|
||||||
|
case "bool":
|
||||||
|
switch v {
|
||||||
|
case "true":
|
||||||
|
return true, nil
|
||||||
|
case "false":
|
||||||
|
return false, nil
|
||||||
|
}
|
||||||
|
return nil, fmt.Errorf("untag: could not parse %q as a boolean", v)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, fmt.Errorf("untag: unrecognized tag type %q", t)
|
||||||
|
}
|
||||||
|
|
||||||
|
func parseTime(v, format string, local bool) (t time.Time, err error) {
|
||||||
|
if local {
|
||||||
|
t, err = time.ParseInLocation(format, v, time.Local)
|
||||||
|
} else {
|
||||||
|
t, err = time.Parse(format, v)
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return time.Time{}, fmt.Errorf("Could not parse %q as a datetime: %w", v, err)
|
||||||
|
}
|
||||||
|
return t, nil
|
||||||
|
}
|
||||||
@@ -0,0 +1,50 @@
|
|||||||
|
// Package testsuite provides helper functions for interoperating with the
|
||||||
|
// language-agnostic TOML test suite at github.com/BurntSushi/toml-test.
|
||||||
|
package testsuite
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"os"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Marshal is a helpfer function for calling toml.Marshal
|
||||||
|
//
|
||||||
|
// Only needed to avoid package import loops.
|
||||||
|
func Marshal(v interface{}) ([]byte, error) {
|
||||||
|
return toml.Marshal(v)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Unmarshal is a helper function for calling toml.Unmarshal.
|
||||||
|
//
|
||||||
|
// Only needed to avoid package import loops.
|
||||||
|
func Unmarshal(data []byte, v interface{}) error {
|
||||||
|
return toml.Unmarshal(data, v)
|
||||||
|
}
|
||||||
|
|
||||||
|
// ValueToTaggedJSON takes a data structure and returns the tagged JSON
|
||||||
|
// representation.
|
||||||
|
func ValueToTaggedJSON(doc interface{}) ([]byte, error) {
|
||||||
|
return json.MarshalIndent(addTag("", doc), "", " ")
|
||||||
|
}
|
||||||
|
|
||||||
|
// DecodeStdin is a helper function for the toml-test binary interface. TOML input
|
||||||
|
// is read from STDIN and a resulting tagged JSON representation is written to
|
||||||
|
// STDOUT.
|
||||||
|
func DecodeStdin() error {
|
||||||
|
var decoded map[string]interface{}
|
||||||
|
|
||||||
|
if err := toml.NewDecoder(os.Stdin).Decode(&decoded); err != nil {
|
||||||
|
return fmt.Errorf("Error decoding TOML: %s", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
j := json.NewEncoder(os.Stdout)
|
||||||
|
j.SetIndent("", " ")
|
||||||
|
if err := j.Encode(addTag("", decoded)); err != nil {
|
||||||
|
return fmt.Errorf("Error encoding JSON: %s", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
+164
-65
@@ -3,6 +3,7 @@ package tracker
|
|||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"sync"
|
||||||
|
|
||||||
"github.com/pelletier/go-toml/v2/internal/ast"
|
"github.com/pelletier/go-toml/v2/internal/ast"
|
||||||
)
|
)
|
||||||
@@ -54,64 +55,104 @@ func (k keyKind) String() string {
|
|||||||
type SeenTracker struct {
|
type SeenTracker struct {
|
||||||
entries []entry
|
entries []entry
|
||||||
currentIdx int
|
currentIdx int
|
||||||
nextID int
|
}
|
||||||
|
|
||||||
|
var pool sync.Pool
|
||||||
|
|
||||||
|
func (s *SeenTracker) reset() {
|
||||||
|
// Always contains a root element at index 0.
|
||||||
|
s.currentIdx = 0
|
||||||
|
if len(s.entries) == 0 {
|
||||||
|
s.entries = make([]entry, 1, 2)
|
||||||
|
} else {
|
||||||
|
s.entries = s.entries[:1]
|
||||||
|
}
|
||||||
|
s.entries[0].child = -1
|
||||||
|
s.entries[0].next = -1
|
||||||
}
|
}
|
||||||
|
|
||||||
type entry struct {
|
type entry struct {
|
||||||
id int
|
// Use -1 to indicate no child or no sibling.
|
||||||
parent int
|
child int
|
||||||
|
next int
|
||||||
|
|
||||||
name []byte
|
name []byte
|
||||||
kind keyKind
|
kind keyKind
|
||||||
explicit bool
|
explicit bool
|
||||||
|
kv bool
|
||||||
}
|
}
|
||||||
|
|
||||||
// Remove all descendent of node at position idx.
|
// Find the index of the child of parentIdx with key k. Returns -1 if
|
||||||
func (s *SeenTracker) clear(idx int) {
|
// it does not exist.
|
||||||
p := s.entries[idx].id
|
func (s *SeenTracker) find(parentIdx int, k []byte) int {
|
||||||
rest := clear(p, s.entries[idx+1:])
|
for i := s.entries[parentIdx].child; i >= 0; i = s.entries[i].next {
|
||||||
s.entries = s.entries[:idx+1+len(rest)]
|
if bytes.Equal(s.entries[i].name, k) {
|
||||||
}
|
return i
|
||||||
|
|
||||||
func clear(parentID int, entries []entry) []entry {
|
|
||||||
for i := 0; i < len(entries); {
|
|
||||||
if entries[i].parent == parentID {
|
|
||||||
id := entries[i].id
|
|
||||||
copy(entries[i:], entries[i+1:])
|
|
||||||
entries = entries[:len(entries)-1]
|
|
||||||
rest := clear(id, entries[i:])
|
|
||||||
entries = entries[:i+len(rest)]
|
|
||||||
} else {
|
|
||||||
i++
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return entries
|
return -1
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *SeenTracker) create(parentIdx int, name []byte, kind keyKind, explicit bool) int {
|
// Remove all descendants of node at position idx.
|
||||||
parentID := s.id(parentIdx)
|
func (s *SeenTracker) clear(idx int) {
|
||||||
|
if idx >= len(s.entries) {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
for i := s.entries[idx].child; i >= 0; {
|
||||||
|
next := s.entries[i].next
|
||||||
|
n := s.entries[0].next
|
||||||
|
s.entries[0].next = i
|
||||||
|
s.entries[i].next = n
|
||||||
|
s.entries[i].name = nil
|
||||||
|
s.clear(i)
|
||||||
|
i = next
|
||||||
|
}
|
||||||
|
|
||||||
|
s.entries[idx].child = -1
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *SeenTracker) create(parentIdx int, name []byte, kind keyKind, explicit bool, kv bool) int {
|
||||||
|
e := entry{
|
||||||
|
child: -1,
|
||||||
|
next: s.entries[parentIdx].child,
|
||||||
|
|
||||||
idx := len(s.entries)
|
|
||||||
s.entries = append(s.entries, entry{
|
|
||||||
id: s.nextID,
|
|
||||||
parent: parentID,
|
|
||||||
name: name,
|
name: name,
|
||||||
kind: kind,
|
kind: kind,
|
||||||
explicit: explicit,
|
explicit: explicit,
|
||||||
})
|
kv: kv,
|
||||||
s.nextID++
|
}
|
||||||
|
var idx int
|
||||||
|
if s.entries[0].next >= 0 {
|
||||||
|
idx = s.entries[0].next
|
||||||
|
s.entries[0].next = s.entries[idx].next
|
||||||
|
s.entries[idx] = e
|
||||||
|
} else {
|
||||||
|
idx = len(s.entries)
|
||||||
|
s.entries = append(s.entries, e)
|
||||||
|
}
|
||||||
|
|
||||||
|
s.entries[parentIdx].child = idx
|
||||||
|
|
||||||
return idx
|
return idx
|
||||||
}
|
}
|
||||||
|
|
||||||
// CheckExpression takes a top-level node and checks that it does not contain keys
|
func (s *SeenTracker) setExplicitFlag(parentIdx int) {
|
||||||
// that have been seen in previous calls, and validates that types are consistent.
|
for i := s.entries[parentIdx].child; i >= 0; i = s.entries[i].next {
|
||||||
|
if s.entries[i].kv {
|
||||||
|
s.entries[i].explicit = true
|
||||||
|
s.entries[i].kv = false
|
||||||
|
}
|
||||||
|
s.setExplicitFlag(i)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// CheckExpression takes a top-level node and checks that it does not contain
|
||||||
|
// keys that have been seen in previous calls, and validates that types are
|
||||||
|
// consistent.
|
||||||
func (s *SeenTracker) CheckExpression(node *ast.Node) error {
|
func (s *SeenTracker) CheckExpression(node *ast.Node) error {
|
||||||
if s.entries == nil {
|
if s.entries == nil {
|
||||||
// s.entries = make([]entry, 0, 8)
|
s.reset()
|
||||||
// Skip ID = 0 to remove the confusion between nodes whose parent has
|
|
||||||
// id 0 and root nodes (parent id is 0 because it's the zero value).
|
|
||||||
s.nextID = 1
|
|
||||||
// Start unscoped, so idx is negative.
|
|
||||||
s.currentIdx = -1
|
|
||||||
}
|
}
|
||||||
switch node.Kind {
|
switch node.Kind {
|
||||||
case ast.KeyValue:
|
case ast.KeyValue:
|
||||||
@@ -126,9 +167,13 @@ func (s *SeenTracker) CheckExpression(node *ast.Node) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (s *SeenTracker) checkTable(node *ast.Node) error {
|
func (s *SeenTracker) checkTable(node *ast.Node) error {
|
||||||
|
if s.currentIdx >= 0 {
|
||||||
|
s.setExplicitFlag(s.currentIdx)
|
||||||
|
}
|
||||||
|
|
||||||
it := node.Key()
|
it := node.Key()
|
||||||
|
|
||||||
parentIdx := -1
|
parentIdx := 0
|
||||||
|
|
||||||
// This code is duplicated in checkArrayTable. This is because factoring
|
// This code is duplicated in checkArrayTable. This is because factoring
|
||||||
// it in a function requires to copy the iterator, or allocate it to the
|
// it in a function requires to copy the iterator, or allocate it to the
|
||||||
@@ -143,7 +188,12 @@ func (s *SeenTracker) checkTable(node *ast.Node) error {
|
|||||||
idx := s.find(parentIdx, k)
|
idx := s.find(parentIdx, k)
|
||||||
|
|
||||||
if idx < 0 {
|
if idx < 0 {
|
||||||
idx = s.create(parentIdx, k, tableKind, false)
|
idx = s.create(parentIdx, k, tableKind, false, false)
|
||||||
|
} else {
|
||||||
|
entry := s.entries[idx]
|
||||||
|
if entry.kind == valueKind {
|
||||||
|
return fmt.Errorf("toml: expected %s to be a table, not a %s", string(k), entry.kind)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
parentIdx = idx
|
parentIdx = idx
|
||||||
}
|
}
|
||||||
@@ -161,7 +211,7 @@ func (s *SeenTracker) checkTable(node *ast.Node) error {
|
|||||||
}
|
}
|
||||||
s.entries[idx].explicit = true
|
s.entries[idx].explicit = true
|
||||||
} else {
|
} else {
|
||||||
idx = s.create(parentIdx, k, tableKind, true)
|
idx = s.create(parentIdx, k, tableKind, true, false)
|
||||||
}
|
}
|
||||||
|
|
||||||
s.currentIdx = idx
|
s.currentIdx = idx
|
||||||
@@ -170,9 +220,13 @@ func (s *SeenTracker) checkTable(node *ast.Node) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (s *SeenTracker) checkArrayTable(node *ast.Node) error {
|
func (s *SeenTracker) checkArrayTable(node *ast.Node) error {
|
||||||
|
if s.currentIdx >= 0 {
|
||||||
|
s.setExplicitFlag(s.currentIdx)
|
||||||
|
}
|
||||||
|
|
||||||
it := node.Key()
|
it := node.Key()
|
||||||
|
|
||||||
parentIdx := -1
|
parentIdx := 0
|
||||||
|
|
||||||
for it.Next() {
|
for it.Next() {
|
||||||
if it.IsLast() {
|
if it.IsLast() {
|
||||||
@@ -184,8 +238,14 @@ func (s *SeenTracker) checkArrayTable(node *ast.Node) error {
|
|||||||
idx := s.find(parentIdx, k)
|
idx := s.find(parentIdx, k)
|
||||||
|
|
||||||
if idx < 0 {
|
if idx < 0 {
|
||||||
idx = s.create(parentIdx, k, tableKind, false)
|
idx = s.create(parentIdx, k, tableKind, false, false)
|
||||||
|
} else {
|
||||||
|
entry := s.entries[idx]
|
||||||
|
if entry.kind == valueKind {
|
||||||
|
return fmt.Errorf("toml: expected %s to be a table, not a %s", string(k), entry.kind)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
parentIdx = idx
|
parentIdx = idx
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -199,7 +259,7 @@ func (s *SeenTracker) checkArrayTable(node *ast.Node) error {
|
|||||||
}
|
}
|
||||||
s.clear(idx)
|
s.clear(idx)
|
||||||
} else {
|
} else {
|
||||||
idx = s.create(parentIdx, k, arrayTableKind, true)
|
idx = s.create(parentIdx, k, arrayTableKind, true, false)
|
||||||
}
|
}
|
||||||
|
|
||||||
s.currentIdx = idx
|
s.currentIdx = idx
|
||||||
@@ -208,50 +268,89 @@ func (s *SeenTracker) checkArrayTable(node *ast.Node) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (s *SeenTracker) checkKeyValue(node *ast.Node) error {
|
func (s *SeenTracker) checkKeyValue(node *ast.Node) error {
|
||||||
it := node.Key()
|
|
||||||
|
|
||||||
parentIdx := s.currentIdx
|
parentIdx := s.currentIdx
|
||||||
|
it := node.Key()
|
||||||
|
|
||||||
for it.Next() {
|
for it.Next() {
|
||||||
k := it.Node().Data
|
k := it.Node().Data
|
||||||
|
|
||||||
idx := s.find(parentIdx, k)
|
idx := s.find(parentIdx, k)
|
||||||
|
|
||||||
if idx >= 0 {
|
if idx < 0 {
|
||||||
if s.entries[idx].kind != tableKind {
|
idx = s.create(parentIdx, k, tableKind, false, true)
|
||||||
return fmt.Errorf("toml: expected %s to be a table, not a %s", string(k), s.entries[idx].kind)
|
|
||||||
}
|
|
||||||
} else {
|
} else {
|
||||||
idx = s.create(parentIdx, k, tableKind, false)
|
entry := s.entries[idx]
|
||||||
|
if it.IsLast() {
|
||||||
|
return fmt.Errorf("toml: key %s is already defined", string(k))
|
||||||
|
} else if entry.kind != tableKind {
|
||||||
|
return fmt.Errorf("toml: expected %s to be a table, not a %s", string(k), entry.kind)
|
||||||
|
} else if entry.explicit {
|
||||||
|
return fmt.Errorf("toml: cannot redefine table %s that has already been explicitly defined", string(k))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
parentIdx = idx
|
parentIdx = idx
|
||||||
}
|
}
|
||||||
|
|
||||||
kind := valueKind
|
s.entries[parentIdx].kind = valueKind
|
||||||
|
|
||||||
if node.Value().Kind == ast.InlineTable {
|
value := node.Value()
|
||||||
kind = tableKind
|
|
||||||
|
switch value.Kind {
|
||||||
|
case ast.InlineTable:
|
||||||
|
return s.checkInlineTable(value)
|
||||||
|
case ast.Array:
|
||||||
|
return s.checkArray(value)
|
||||||
}
|
}
|
||||||
s.entries[parentIdx].kind = kind
|
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *SeenTracker) id(idx int) int {
|
func (s *SeenTracker) checkArray(node *ast.Node) error {
|
||||||
if idx >= 0 {
|
it := node.Children()
|
||||||
return s.entries[idx].id
|
for it.Next() {
|
||||||
|
n := it.Node()
|
||||||
|
switch n.Kind {
|
||||||
|
case ast.InlineTable:
|
||||||
|
err := s.checkInlineTable(n)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
case ast.Array:
|
||||||
|
err := s.checkArray(n)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return 0
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *SeenTracker) find(parentIdx int, k []byte) int {
|
func (s *SeenTracker) checkInlineTable(node *ast.Node) error {
|
||||||
parentID := s.id(parentIdx)
|
if pool.New == nil {
|
||||||
|
pool.New = func() interface{} {
|
||||||
for i := parentIdx + 1; i < len(s.entries); i++ {
|
return &SeenTracker{}
|
||||||
if s.entries[i].parent == parentID && bytes.Equal(s.entries[i].name, k) {
|
|
||||||
return i
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return -1
|
s = pool.Get().(*SeenTracker)
|
||||||
|
s.reset()
|
||||||
|
|
||||||
|
it := node.Children()
|
||||||
|
for it.Next() {
|
||||||
|
n := it.Node()
|
||||||
|
err := s.checkKeyValue(n)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// As inline tables are self-contained, the tracker does not
|
||||||
|
// need to retain the details of what they contain. The
|
||||||
|
// keyValue element that creates the inline table is kept to
|
||||||
|
// mark the presence of the inline table and prevent
|
||||||
|
// redefinition of its keys: check* functions cannot walk into
|
||||||
|
// a value.
|
||||||
|
pool.Put(s)
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -0,0 +1,16 @@
|
|||||||
|
package tracker
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
"unsafe"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestEntrySize(t *testing.T) {
|
||||||
|
// Validate no regression on the size of entry{}. This is a critical bit for
|
||||||
|
// performance of unmarshaling documents. Should only be increased with care
|
||||||
|
// and a very good reason.
|
||||||
|
maxExpectedEntrySize := 48
|
||||||
|
require.LessOrEqual(t, int(unsafe.Sizeof(entry{})), maxExpectedEntrySize)
|
||||||
|
}
|
||||||
+79
-259
@@ -1,300 +1,120 @@
|
|||||||
// Implementation of TOML's local date/time.
|
|
||||||
// Copied over from https://github.com/googleapis/google-cloud-go/blob/master/civil/civil.go
|
|
||||||
// to avoid pulling all the Google dependencies.
|
|
||||||
//
|
|
||||||
// Copyright 2016 Google LLC
|
|
||||||
//
|
|
||||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
// you may not use this file except in compliance with the License.
|
|
||||||
// You may obtain a copy of the License at
|
|
||||||
//
|
|
||||||
// http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
//
|
|
||||||
// Unless required by applicable law or agreed to in writing, software
|
|
||||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
// See the License for the specific language governing permissions and
|
|
||||||
// limitations under the License.
|
|
||||||
|
|
||||||
// Package civil implements types for civil time, a time-zone-independent
|
|
||||||
// representation of time that follows the rules of the proleptic
|
|
||||||
// Gregorian calendar with exactly 24-hour days, 60-minute hours, and 60-second
|
|
||||||
// minutes.
|
|
||||||
//
|
|
||||||
// Because they lack location information, these types do not represent unique
|
|
||||||
// moments or intervals of time. Use time.Time for that purpose.
|
|
||||||
|
|
||||||
package toml
|
package toml
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
// A LocalDate represents a date (year, month, day).
|
// LocalDate represents a calendar day in no specific timezone.
|
||||||
//
|
|
||||||
// This type does not include location information, and therefore does not
|
|
||||||
// describe a unique 24-hour timespan.
|
|
||||||
type LocalDate struct {
|
type LocalDate struct {
|
||||||
Year int // Year (e.g., 2014).
|
Year int
|
||||||
Month time.Month // Month of the year (January = 1, ...).
|
Month int
|
||||||
Day int // Day of the month, starting at 1.
|
Day int
|
||||||
}
|
}
|
||||||
|
|
||||||
// LocalDateOf returns the LocalDate in which a time occurs in that time's location.
|
// AsTime converts d into a specific time instance at midnight in zone.
|
||||||
func LocalDateOf(t time.Time) LocalDate {
|
func (d LocalDate) AsTime(zone *time.Location) time.Time {
|
||||||
var d LocalDate
|
return time.Date(d.Year, time.Month(d.Month), d.Day, 0, 0, 0, 0, zone)
|
||||||
d.Year, d.Month, d.Day = t.Date()
|
|
||||||
|
|
||||||
return d
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// ParseLocalDate parses a string in RFC3339 full-date format and returns the date value it represents.
|
// String returns RFC 3339 representation of d.
|
||||||
func ParseLocalDate(s string) (LocalDate, error) {
|
|
||||||
t, err := time.Parse("2006-01-02", s)
|
|
||||||
if err != nil {
|
|
||||||
return LocalDate{}, err
|
|
||||||
}
|
|
||||||
|
|
||||||
return LocalDateOf(t), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// String returns the date in RFC3339 full-date format.
|
|
||||||
func (d LocalDate) String() string {
|
func (d LocalDate) String() string {
|
||||||
return fmt.Sprintf("%04d-%02d-%02d", d.Year, d.Month, d.Day)
|
return fmt.Sprintf("%04d-%02d-%02d", d.Year, d.Month, d.Day)
|
||||||
}
|
}
|
||||||
|
|
||||||
// IsValid reports whether the date is valid.
|
// MarshalText returns RFC 3339 representation of d.
|
||||||
func (d LocalDate) IsValid() bool {
|
|
||||||
return LocalDateOf(d.In(time.UTC)) == d
|
|
||||||
}
|
|
||||||
|
|
||||||
// In returns the time corresponding to time 00:00:00 of the date in the location.
|
|
||||||
//
|
|
||||||
// In is always consistent with time.LocalDate, even when time.LocalDate returns a time
|
|
||||||
// on a different day. For example, if loc is America/Indiana/Vincennes, then both
|
|
||||||
// time.LocalDate(1955, time.May, 1, 0, 0, 0, 0, loc)
|
|
||||||
// and
|
|
||||||
// civil.LocalDate{Year: 1955, Month: time.May, Day: 1}.In(loc)
|
|
||||||
// return 23:00:00 on April 30, 1955.
|
|
||||||
//
|
|
||||||
// In panics if loc is nil.
|
|
||||||
func (d LocalDate) In(loc *time.Location) time.Time {
|
|
||||||
return time.Date(d.Year, d.Month, d.Day, 0, 0, 0, 0, loc)
|
|
||||||
}
|
|
||||||
|
|
||||||
// AddDays returns the date that is n days in the future.
|
|
||||||
// n can also be negative to go into the past.
|
|
||||||
func (d LocalDate) AddDays(n int) LocalDate {
|
|
||||||
return LocalDateOf(d.In(time.UTC).AddDate(0, 0, n))
|
|
||||||
}
|
|
||||||
|
|
||||||
// DaysSince returns the signed number of days between the date and s, not including the end day.
|
|
||||||
// This is the inverse operation to AddDays.
|
|
||||||
func (d LocalDate) DaysSince(s LocalDate) (days int) {
|
|
||||||
// We convert to Unix time so we do not have to worry about leap seconds:
|
|
||||||
// Unix time increases by exactly 86400 seconds per day.
|
|
||||||
deltaUnix := d.In(time.UTC).Unix() - s.In(time.UTC).Unix()
|
|
||||||
|
|
||||||
const secondsInADay = 86400
|
|
||||||
|
|
||||||
return int(deltaUnix / secondsInADay)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Before reports whether d1 occurs before future date.
|
|
||||||
func (d LocalDate) Before(future LocalDate) bool {
|
|
||||||
if d.Year != future.Year {
|
|
||||||
return d.Year < future.Year
|
|
||||||
}
|
|
||||||
|
|
||||||
if d.Month != future.Month {
|
|
||||||
return d.Month < future.Month
|
|
||||||
}
|
|
||||||
|
|
||||||
return d.Day < future.Day
|
|
||||||
}
|
|
||||||
|
|
||||||
// After reports whether d1 occurs after past date.
|
|
||||||
func (d LocalDate) After(past LocalDate) bool {
|
|
||||||
return past.Before(d)
|
|
||||||
}
|
|
||||||
|
|
||||||
// MarshalText implements the encoding.TextMarshaler interface.
|
|
||||||
// The output is the result of d.String().
|
|
||||||
func (d LocalDate) MarshalText() ([]byte, error) {
|
func (d LocalDate) MarshalText() ([]byte, error) {
|
||||||
return []byte(d.String()), nil
|
return []byte(d.String()), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// UnmarshalText implements the encoding.TextUnmarshaler interface.
|
// UnmarshalText parses b using RFC 3339 to fill d.
|
||||||
// The date is expected to be a string in a format accepted by ParseLocalDate.
|
func (d *LocalDate) UnmarshalText(b []byte) error {
|
||||||
func (d *LocalDate) UnmarshalText(data []byte) error {
|
res, err := parseLocalDate(b)
|
||||||
var err error
|
if err != nil {
|
||||||
*d, err = ParseLocalDate(string(data))
|
return err
|
||||||
|
}
|
||||||
return err
|
*d = res
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// A LocalTime represents a time with nanosecond precision.
|
// LocalTime represents a time of day of no specific day in no specific
|
||||||
//
|
// timezone.
|
||||||
// This type does not include location information, and therefore does not
|
|
||||||
// describe a unique moment in time.
|
|
||||||
//
|
|
||||||
// This type exists to represent the TIME type in storage-based APIs like BigQuery.
|
|
||||||
// Most operations on Times are unlikely to be meaningful. Prefer the LocalDateTime type.
|
|
||||||
type LocalTime struct {
|
type LocalTime struct {
|
||||||
Hour int // The hour of the day in 24-hour format; range [0-23]
|
Hour int // Hour of the day: [0; 24[
|
||||||
Minute int // The minute of the hour; range [0-59]
|
Minute int // Minute of the hour: [0; 60[
|
||||||
Second int // The second of the minute; range [0-59]
|
Second int // Second of the minute: [0; 60[
|
||||||
Nanosecond int // The nanosecond of the second; range [0-999999999]
|
Nanosecond int // Nanoseconds within the second: [0, 1000000000[
|
||||||
|
Precision int // Number of digits to display for Nanosecond.
|
||||||
}
|
}
|
||||||
|
|
||||||
// LocalTimeOf returns the LocalTime representing the time of day in which a time occurs
|
// String returns RFC 3339 representation of d.
|
||||||
// in that time's location. It ignores the date.
|
// If d.Nanosecond and d.Precision are zero, the time won't have a nanosecond
|
||||||
func LocalTimeOf(t time.Time) LocalTime {
|
// component. If d.Nanosecond > 0 but d.Precision = 0, then the minimum number
|
||||||
var tm LocalTime
|
// of digits for nanoseconds is provided.
|
||||||
tm.Hour, tm.Minute, tm.Second = t.Clock()
|
func (d LocalTime) String() string {
|
||||||
tm.Nanosecond = t.Nanosecond()
|
s := fmt.Sprintf("%02d:%02d:%02d", d.Hour, d.Minute, d.Second)
|
||||||
|
|
||||||
return tm
|
if d.Precision > 0 {
|
||||||
|
s += fmt.Sprintf(".%09d", d.Nanosecond)[:d.Precision+1]
|
||||||
|
} else if d.Nanosecond > 0 {
|
||||||
|
// Nanoseconds are specified, but precision is not provided. Use the
|
||||||
|
// minimum.
|
||||||
|
s += strings.Trim(fmt.Sprintf(".%09d", d.Nanosecond), "0")
|
||||||
|
}
|
||||||
|
|
||||||
|
return s
|
||||||
}
|
}
|
||||||
|
|
||||||
// ParseLocalTime parses a string and returns the time value it represents.
|
// MarshalText returns RFC 3339 representation of d.
|
||||||
// ParseLocalTime accepts an extended form of the RFC3339 partial-time format. After
|
func (d LocalTime) MarshalText() ([]byte, error) {
|
||||||
// the HH:MM:SS part of the string, an optional fractional part may appear,
|
return []byte(d.String()), nil
|
||||||
// consisting of a decimal point followed by one to nine decimal digits.
|
}
|
||||||
// (RFC3339 admits only one digit after the decimal point).
|
|
||||||
func ParseLocalTime(s string) (LocalTime, error) {
|
// UnmarshalText parses b using RFC 3339 to fill d.
|
||||||
t, err := time.Parse("15:04:05.999999999", s)
|
func (d *LocalTime) UnmarshalText(b []byte) error {
|
||||||
|
res, left, err := parseLocalTime(b)
|
||||||
|
if err == nil && len(left) != 0 {
|
||||||
|
err = newDecodeError(left, "extra characters")
|
||||||
|
}
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return LocalTime{}, err
|
return err
|
||||||
}
|
}
|
||||||
|
*d = res
|
||||||
return LocalTimeOf(t), nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// String returns the date in the format described in ParseLocalTime. If Nanoseconds
|
// LocalDateTime represents a time of a specific day in no specific timezone.
|
||||||
// is zero, no fractional part will be generated. Otherwise, the result will
|
|
||||||
// end with a fractional part consisting of a decimal point and nine digits.
|
|
||||||
func (t LocalTime) String() string {
|
|
||||||
s := fmt.Sprintf("%02d:%02d:%02d", t.Hour, t.Minute, t.Second)
|
|
||||||
if t.Nanosecond == 0 {
|
|
||||||
return s
|
|
||||||
}
|
|
||||||
|
|
||||||
return s + fmt.Sprintf(".%09d", t.Nanosecond)
|
|
||||||
}
|
|
||||||
|
|
||||||
// IsValid reports whether the time is valid.
|
|
||||||
func (t LocalTime) IsValid() bool {
|
|
||||||
// Construct a non-zero time.
|
|
||||||
tm := time.Date(2, 2, 2, t.Hour, t.Minute, t.Second, t.Nanosecond, time.UTC)
|
|
||||||
|
|
||||||
return LocalTimeOf(tm) == t
|
|
||||||
}
|
|
||||||
|
|
||||||
// MarshalText implements the encoding.TextMarshaler interface.
|
|
||||||
// The output is the result of t.String().
|
|
||||||
func (t LocalTime) MarshalText() ([]byte, error) {
|
|
||||||
return []byte(t.String()), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// UnmarshalText implements the encoding.TextUnmarshaler interface.
|
|
||||||
// The time is expected to be a string in a format accepted by ParseLocalTime.
|
|
||||||
func (t *LocalTime) UnmarshalText(data []byte) error {
|
|
||||||
var err error
|
|
||||||
*t, err = ParseLocalTime(string(data))
|
|
||||||
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
// A LocalDateTime represents a date and time.
|
|
||||||
//
|
|
||||||
// This type does not include location information, and therefore does not
|
|
||||||
// describe a unique moment in time.
|
|
||||||
type LocalDateTime struct {
|
type LocalDateTime struct {
|
||||||
Date LocalDate
|
LocalDate
|
||||||
Time LocalTime
|
LocalTime
|
||||||
}
|
}
|
||||||
|
|
||||||
// Note: We deliberately do not embed LocalDate into LocalDateTime, to avoid promoting AddDays and Sub.
|
// AsTime converts d into a specific time instance in zone.
|
||||||
|
func (d LocalDateTime) AsTime(zone *time.Location) time.Time {
|
||||||
|
return time.Date(d.Year, time.Month(d.Month), d.Day, d.Hour, d.Minute, d.Second, d.Nanosecond, zone)
|
||||||
|
}
|
||||||
|
|
||||||
// LocalDateTimeOf returns the LocalDateTime in which a time occurs in that time's location.
|
// String returns RFC 3339 representation of d.
|
||||||
func LocalDateTimeOf(t time.Time) LocalDateTime {
|
func (d LocalDateTime) String() string {
|
||||||
return LocalDateTime{
|
return d.LocalDate.String() + "T" + d.LocalTime.String()
|
||||||
Date: LocalDateOf(t),
|
}
|
||||||
Time: LocalTimeOf(t),
|
|
||||||
|
// MarshalText returns RFC 3339 representation of d.
|
||||||
|
func (d LocalDateTime) MarshalText() ([]byte, error) {
|
||||||
|
return []byte(d.String()), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// UnmarshalText parses b using RFC 3339 to fill d.
|
||||||
|
func (d *LocalDateTime) UnmarshalText(data []byte) error {
|
||||||
|
res, left, err := parseLocalDateTime(data)
|
||||||
|
if err == nil && len(left) != 0 {
|
||||||
|
err = newDecodeError(left, "extra characters")
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
// ParseLocalDateTime parses a string and returns the LocalDateTime it represents.
|
|
||||||
// ParseLocalDateTime accepts a variant of the RFC3339 date-time format that omits
|
|
||||||
// the time offset but includes an optional fractional time, as described in
|
|
||||||
// ParseLocalTime. Informally, the accepted format is
|
|
||||||
// YYYY-MM-DDTHH:MM:SS[.FFFFFFFFF]
|
|
||||||
// where the 'T' may be a lower-case 't'.
|
|
||||||
func ParseLocalDateTime(s string) (LocalDateTime, error) {
|
|
||||||
t, err := time.Parse("2006-01-02T15:04:05.999999999", s)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t, err = time.Parse("2006-01-02t15:04:05.999999999", s)
|
return err
|
||||||
if err != nil {
|
|
||||||
return LocalDateTime{}, err
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return LocalDateTimeOf(t), nil
|
*d = res
|
||||||
}
|
return nil
|
||||||
|
|
||||||
// String returns the date in the format described in ParseLocalDate.
|
|
||||||
func (dt LocalDateTime) String() string {
|
|
||||||
return dt.Date.String() + "T" + dt.Time.String()
|
|
||||||
}
|
|
||||||
|
|
||||||
// IsValid reports whether the datetime is valid.
|
|
||||||
func (dt LocalDateTime) IsValid() bool {
|
|
||||||
return dt.Date.IsValid() && dt.Time.IsValid()
|
|
||||||
}
|
|
||||||
|
|
||||||
// In returns the time corresponding to the LocalDateTime in the given location.
|
|
||||||
//
|
|
||||||
// If the time is missing or ambigous at the location, In returns the same
|
|
||||||
// result as time.LocalDate. For example, if loc is America/Indiana/Vincennes, then
|
|
||||||
// both
|
|
||||||
// time.LocalDate(1955, time.May, 1, 0, 30, 0, 0, loc)
|
|
||||||
// and
|
|
||||||
// civil.LocalDateTime{
|
|
||||||
// civil.LocalDate{Year: 1955, Month: time.May, Day: 1}},
|
|
||||||
// civil.LocalTime{Minute: 30}}.In(loc)
|
|
||||||
// return 23:30:00 on April 30, 1955.
|
|
||||||
//
|
|
||||||
// In panics if loc is nil.
|
|
||||||
func (dt LocalDateTime) In(loc *time.Location) time.Time {
|
|
||||||
return time.Date(
|
|
||||||
dt.Date.Year, dt.Date.Month, dt.Date.Day,
|
|
||||||
dt.Time.Hour, dt.Time.Minute, dt.Time.Second, dt.Time.Nanosecond, loc,
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Before reports whether dt occurs before future.
|
|
||||||
func (dt LocalDateTime) Before(future LocalDateTime) bool {
|
|
||||||
return dt.In(time.UTC).Before(future.In(time.UTC))
|
|
||||||
}
|
|
||||||
|
|
||||||
// After reports whether dt occurs after past.
|
|
||||||
func (dt LocalDateTime) After(past LocalDateTime) bool {
|
|
||||||
return past.Before(dt)
|
|
||||||
}
|
|
||||||
|
|
||||||
// MarshalText implements the encoding.TextMarshaler interface.
|
|
||||||
// The output is the result of dt.String().
|
|
||||||
func (dt LocalDateTime) MarshalText() ([]byte, error) {
|
|
||||||
return []byte(dt.String()), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// UnmarshalText implements the encoding.TextUnmarshaler interface.
|
|
||||||
// The datetime is expected to be a string in a format accepted by ParseLocalDateTime.
|
|
||||||
func (dt *LocalDateTime) UnmarshalText(data []byte) error {
|
|
||||||
var err error
|
|
||||||
*dt, err = ParseLocalDateTime(string(data))
|
|
||||||
|
|
||||||
return err
|
|
||||||
}
|
}
|
||||||
|
|||||||
+83
-454
@@ -1,489 +1,118 @@
|
|||||||
// Copyright 2016 Google LLC
|
package toml_test
|
||||||
//
|
|
||||||
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
||||||
// you may not use this file except in compliance with the License.
|
|
||||||
// You may obtain a copy of the License at
|
|
||||||
//
|
|
||||||
// http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
//
|
|
||||||
// Unless required by applicable law or agreed to in writing, software
|
|
||||||
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
||||||
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
||||||
// See the License for the specific language governing permissions and
|
|
||||||
// limitations under the License.
|
|
||||||
|
|
||||||
package toml
|
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
|
||||||
"reflect"
|
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
func cmpEqual(x, y interface{}) bool {
|
func TestLocalDate_AsTime(t *testing.T) {
|
||||||
return reflect.DeepEqual(x, y)
|
d := toml.LocalDate{2021, 6, 8}
|
||||||
|
cast := d.AsTime(time.UTC)
|
||||||
|
require.Equal(t, time.Date(2021, time.June, 8, 0, 0, 0, 0, time.UTC), cast)
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDates(t *testing.T) {
|
func TestLocalDate_String(t *testing.T) {
|
||||||
|
d := toml.LocalDate{2021, 6, 8}
|
||||||
for _, test := range []struct {
|
require.Equal(t, "2021-06-08", d.String())
|
||||||
date LocalDate
|
|
||||||
loc *time.Location
|
|
||||||
wantStr string
|
|
||||||
wantTime time.Time
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
date: LocalDate{2014, 7, 29},
|
|
||||||
loc: time.Local,
|
|
||||||
wantStr: "2014-07-29",
|
|
||||||
wantTime: time.Date(2014, time.July, 29, 0, 0, 0, 0, time.Local),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
date: LocalDateOf(time.Date(2014, 8, 20, 15, 8, 43, 1, time.Local)),
|
|
||||||
loc: time.UTC,
|
|
||||||
wantStr: "2014-08-20",
|
|
||||||
wantTime: time.Date(2014, 8, 20, 0, 0, 0, 0, time.UTC),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
date: LocalDateOf(time.Date(999, time.January, 26, 0, 0, 0, 0, time.Local)),
|
|
||||||
loc: time.UTC,
|
|
||||||
wantStr: "0999-01-26",
|
|
||||||
wantTime: time.Date(999, 1, 26, 0, 0, 0, 0, time.UTC),
|
|
||||||
},
|
|
||||||
} {
|
|
||||||
if got := test.date.String(); got != test.wantStr {
|
|
||||||
t.Errorf("%#v.String() = %q, want %q", test.date, got, test.wantStr)
|
|
||||||
}
|
|
||||||
|
|
||||||
if got := test.date.In(test.loc); !got.Equal(test.wantTime) {
|
|
||||||
t.Errorf("%#v.In(%v) = %v, want %v", test.date, test.loc, got, test.wantTime)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDateIsValid(t *testing.T) {
|
func TestLocalDate_MarshalText(t *testing.T) {
|
||||||
|
d := toml.LocalDate{2021, 6, 8}
|
||||||
for _, test := range []struct {
|
b, err := d.MarshalText()
|
||||||
date LocalDate
|
require.NoError(t, err)
|
||||||
want bool
|
require.Equal(t, []byte("2021-06-08"), b)
|
||||||
}{
|
|
||||||
{LocalDate{2014, 7, 29}, true},
|
|
||||||
{LocalDate{2000, 2, 29}, true},
|
|
||||||
{LocalDate{10000, 12, 31}, true},
|
|
||||||
{LocalDate{1, 1, 1}, true},
|
|
||||||
{LocalDate{0, 1, 1}, true}, // year zero is OK
|
|
||||||
{LocalDate{-1, 1, 1}, true}, // negative year is OK
|
|
||||||
{LocalDate{1, 0, 1}, false},
|
|
||||||
{LocalDate{1, 1, 0}, false},
|
|
||||||
{LocalDate{2016, 1, 32}, false},
|
|
||||||
{LocalDate{2016, 13, 1}, false},
|
|
||||||
{LocalDate{1, -1, 1}, false},
|
|
||||||
{LocalDate{1, 1, -1}, false},
|
|
||||||
} {
|
|
||||||
got := test.date.IsValid()
|
|
||||||
if got != test.want {
|
|
||||||
t.Errorf("%#v: got %t, want %t", test.date, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestParseDate(t *testing.T) {
|
func TestLocalDate_UnmarshalMarshalText(t *testing.T) {
|
||||||
|
d := toml.LocalDate{}
|
||||||
|
err := d.UnmarshalText([]byte("2021-06-08"))
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, toml.LocalDate{2021, 6, 8}, d)
|
||||||
|
|
||||||
var emptyDate LocalDate
|
err = d.UnmarshalText([]byte("what"))
|
||||||
|
require.Error(t, err)
|
||||||
for _, test := range []struct {
|
|
||||||
str string
|
|
||||||
want LocalDate // if empty, expect an error
|
|
||||||
}{
|
|
||||||
{"2016-01-02", LocalDate{2016, 1, 2}},
|
|
||||||
{"2016-12-31", LocalDate{2016, 12, 31}},
|
|
||||||
{"0003-02-04", LocalDate{3, 2, 4}},
|
|
||||||
{"999-01-26", emptyDate},
|
|
||||||
{"", emptyDate},
|
|
||||||
{"2016-01-02x", emptyDate},
|
|
||||||
} {
|
|
||||||
got, err := ParseLocalDate(test.str)
|
|
||||||
if got != test.want {
|
|
||||||
t.Errorf("ParseLocalDate(%q) = %+v, want %+v", test.str, got, test.want)
|
|
||||||
}
|
|
||||||
|
|
||||||
if err != nil && test.want != (emptyDate) {
|
|
||||||
t.Errorf("Unexpected error %v from ParseLocalDate(%q)", err, test.str)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDateArithmetic(t *testing.T) {
|
func TestLocalTime_String(t *testing.T) {
|
||||||
|
d := toml.LocalTime{20, 12, 1, 2, 9}
|
||||||
for _, test := range []struct {
|
require.Equal(t, "20:12:01.000000002", d.String())
|
||||||
desc string
|
d = toml.LocalTime{20, 12, 1, 0, 0}
|
||||||
start LocalDate
|
require.Equal(t, "20:12:01", d.String())
|
||||||
end LocalDate
|
d = toml.LocalTime{20, 12, 1, 0, 9}
|
||||||
days int
|
require.Equal(t, "20:12:01.000000000", d.String())
|
||||||
}{
|
d = toml.LocalTime{20, 12, 1, 100, 0}
|
||||||
{
|
require.Equal(t, "20:12:01.0000001", d.String())
|
||||||
desc: "zero days noop",
|
|
||||||
start: LocalDate{2014, 5, 9},
|
|
||||||
end: LocalDate{2014, 5, 9},
|
|
||||||
days: 0,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
desc: "crossing a year boundary",
|
|
||||||
start: LocalDate{2014, 12, 31},
|
|
||||||
end: LocalDate{2015, 1, 1},
|
|
||||||
days: 1,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
desc: "negative number of days",
|
|
||||||
start: LocalDate{2015, 1, 1},
|
|
||||||
end: LocalDate{2014, 12, 31},
|
|
||||||
days: -1,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
desc: "full leap year",
|
|
||||||
start: LocalDate{2004, 1, 1},
|
|
||||||
end: LocalDate{2005, 1, 1},
|
|
||||||
days: 366,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
desc: "full non-leap year",
|
|
||||||
start: LocalDate{2001, 1, 1},
|
|
||||||
end: LocalDate{2002, 1, 1},
|
|
||||||
days: 365,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
desc: "crossing a leap second",
|
|
||||||
start: LocalDate{1972, 6, 30},
|
|
||||||
end: LocalDate{1972, 7, 1},
|
|
||||||
days: 1,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
desc: "dates before the unix epoch",
|
|
||||||
start: LocalDate{101, 1, 1},
|
|
||||||
end: LocalDate{102, 1, 1},
|
|
||||||
days: 365,
|
|
||||||
},
|
|
||||||
} {
|
|
||||||
if got := test.start.AddDays(test.days); got != test.end {
|
|
||||||
t.Errorf("[%s] %#v.AddDays(%v) = %#v, want %#v", test.desc, test.start, test.days, got, test.end)
|
|
||||||
}
|
|
||||||
|
|
||||||
if got := test.end.DaysSince(test.start); got != test.days {
|
|
||||||
t.Errorf("[%s] %#v.Sub(%#v) = %v, want %v", test.desc, test.end, test.start, got, test.days)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDateBefore(t *testing.T) {
|
func TestLocalTime_MarshalText(t *testing.T) {
|
||||||
|
d := toml.LocalTime{20, 12, 1, 2, 9}
|
||||||
for _, test := range []struct {
|
b, err := d.MarshalText()
|
||||||
d1, d2 LocalDate
|
require.NoError(t, err)
|
||||||
want bool
|
require.Equal(t, []byte("20:12:01.000000002"), b)
|
||||||
}{
|
|
||||||
{LocalDate{2016, 12, 31}, LocalDate{2017, 1, 1}, true},
|
|
||||||
{LocalDate{2016, 1, 1}, LocalDate{2016, 1, 1}, false},
|
|
||||||
{LocalDate{2016, 12, 30}, LocalDate{2016, 12, 31}, true},
|
|
||||||
{LocalDate{2016, 1, 30}, LocalDate{2016, 12, 31}, true},
|
|
||||||
} {
|
|
||||||
if got := test.d1.Before(test.d2); got != test.want {
|
|
||||||
t.Errorf("%v.Before(%v): got %t, want %t", test.d1, test.d2, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDateAfter(t *testing.T) {
|
func TestLocalTime_UnmarshalMarshalText(t *testing.T) {
|
||||||
|
d := toml.LocalTime{}
|
||||||
|
err := d.UnmarshalText([]byte("20:12:01.000000002"))
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, toml.LocalTime{20, 12, 1, 2, 9}, d)
|
||||||
|
|
||||||
for _, test := range []struct {
|
err = d.UnmarshalText([]byte("what"))
|
||||||
d1, d2 LocalDate
|
require.Error(t, err)
|
||||||
want bool
|
|
||||||
}{
|
err = d.UnmarshalText([]byte("20:12:01.000000002 bad"))
|
||||||
{LocalDate{2016, 12, 31}, LocalDate{2017, 1, 1}, false},
|
require.Error(t, err)
|
||||||
{LocalDate{2016, 1, 1}, LocalDate{2016, 1, 1}, false},
|
|
||||||
{LocalDate{2016, 12, 30}, LocalDate{2016, 12, 31}, false},
|
|
||||||
} {
|
|
||||||
if got := test.d1.After(test.d2); got != test.want {
|
|
||||||
t.Errorf("%v.After(%v): got %t, want %t", test.d1, test.d2, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestTimeToString(t *testing.T) {
|
func TestLocalTime_RoundTrip(t *testing.T) {
|
||||||
|
var d struct{ A toml.LocalTime }
|
||||||
for _, test := range []struct {
|
err := toml.Unmarshal([]byte("a=20:12:01.500"), &d)
|
||||||
str string
|
require.NoError(t, err)
|
||||||
time LocalTime
|
require.Equal(t, "20:12:01.500", d.A.String())
|
||||||
roundTrip bool // ParseLocalTime(str).String() == str?
|
|
||||||
}{
|
|
||||||
{"13:26:33", LocalTime{13, 26, 33, 0}, true},
|
|
||||||
{"01:02:03.000023456", LocalTime{1, 2, 3, 23456}, true},
|
|
||||||
{"00:00:00.000000001", LocalTime{0, 0, 0, 1}, true},
|
|
||||||
{"13:26:03.1", LocalTime{13, 26, 3, 100000000}, false},
|
|
||||||
{"13:26:33.0000003", LocalTime{13, 26, 33, 300}, false},
|
|
||||||
} {
|
|
||||||
gotTime, err := ParseLocalTime(test.str)
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("ParseLocalTime(%q): got error: %v", test.str, err)
|
|
||||||
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
if gotTime != test.time {
|
|
||||||
t.Errorf("ParseLocalTime(%q) = %+v, want %+v", test.str, gotTime, test.time)
|
|
||||||
}
|
|
||||||
|
|
||||||
if test.roundTrip {
|
|
||||||
gotStr := test.time.String()
|
|
||||||
if gotStr != test.str {
|
|
||||||
t.Errorf("%#v.String() = %q, want %q", test.time, gotStr, test.str)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestTimeOf(t *testing.T) {
|
func TestLocalDateTime_AsTime(t *testing.T) {
|
||||||
|
d := toml.LocalDateTime{
|
||||||
for _, test := range []struct {
|
toml.LocalDate{2021, 6, 8},
|
||||||
time time.Time
|
toml.LocalTime{20, 12, 1, 2, 9},
|
||||||
want LocalTime
|
|
||||||
}{
|
|
||||||
{time.Date(2014, 8, 20, 15, 8, 43, 1, time.Local), LocalTime{15, 8, 43, 1}},
|
|
||||||
{time.Date(1, 1, 1, 0, 0, 0, 0, time.UTC), LocalTime{0, 0, 0, 0}},
|
|
||||||
} {
|
|
||||||
if got := LocalTimeOf(test.time); got != test.want {
|
|
||||||
t.Errorf("LocalTimeOf(%v) = %+v, want %+v", test.time, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
cast := d.AsTime(time.UTC)
|
||||||
|
require.Equal(t, time.Date(2021, time.June, 8, 20, 12, 1, 2, time.UTC), cast)
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestTimeIsValid(t *testing.T) {
|
func TestLocalDateTime_String(t *testing.T) {
|
||||||
|
d := toml.LocalDateTime{
|
||||||
for _, test := range []struct {
|
toml.LocalDate{2021, 6, 8},
|
||||||
time LocalTime
|
toml.LocalTime{20, 12, 1, 2, 9},
|
||||||
want bool
|
|
||||||
}{
|
|
||||||
{LocalTime{0, 0, 0, 0}, true},
|
|
||||||
{LocalTime{23, 0, 0, 0}, true},
|
|
||||||
{LocalTime{23, 59, 59, 999999999}, true},
|
|
||||||
{LocalTime{24, 59, 59, 999999999}, false},
|
|
||||||
{LocalTime{23, 60, 59, 999999999}, false},
|
|
||||||
{LocalTime{23, 59, 60, 999999999}, false},
|
|
||||||
{LocalTime{23, 59, 59, 1000000000}, false},
|
|
||||||
{LocalTime{-1, 0, 0, 0}, false},
|
|
||||||
{LocalTime{0, -1, 0, 0}, false},
|
|
||||||
{LocalTime{0, 0, -1, 0}, false},
|
|
||||||
{LocalTime{0, 0, 0, -1}, false},
|
|
||||||
} {
|
|
||||||
got := test.time.IsValid()
|
|
||||||
if got != test.want {
|
|
||||||
t.Errorf("%#v: got %t, want %t", test.time, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
require.Equal(t, "2021-06-08T20:12:01.000000002", d.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestDateTimeToString(t *testing.T) {
|
func TestLocalDateTime_MarshalText(t *testing.T) {
|
||||||
|
d := toml.LocalDateTime{
|
||||||
for _, test := range []struct {
|
toml.LocalDate{2021, 6, 8},
|
||||||
str string
|
toml.LocalTime{20, 12, 1, 2, 9},
|
||||||
dateTime LocalDateTime
|
|
||||||
roundTrip bool // ParseLocalDateTime(str).String() == str?
|
|
||||||
}{
|
|
||||||
{"2016-03-22T13:26:33", LocalDateTime{LocalDate{2016, 3, 22}, LocalTime{13, 26, 33, 0}}, true},
|
|
||||||
{"2016-03-22T13:26:33.000000600", LocalDateTime{LocalDate{2016, 3, 22}, LocalTime{13, 26, 33, 600}}, true},
|
|
||||||
{"2016-03-22t13:26:33", LocalDateTime{LocalDate{2016, 3, 22}, LocalTime{13, 26, 33, 0}}, false},
|
|
||||||
} {
|
|
||||||
gotDateTime, err := ParseLocalDateTime(test.str)
|
|
||||||
if err != nil {
|
|
||||||
t.Errorf("ParseLocalDateTime(%q): got error: %v", test.str, err)
|
|
||||||
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
if gotDateTime != test.dateTime {
|
|
||||||
t.Errorf("ParseLocalDateTime(%q) = %+v, want %+v", test.str, gotDateTime, test.dateTime)
|
|
||||||
}
|
|
||||||
|
|
||||||
if test.roundTrip {
|
|
||||||
gotStr := test.dateTime.String()
|
|
||||||
if gotStr != test.str {
|
|
||||||
t.Errorf("%#v.String() = %q, want %q", test.dateTime, gotStr, test.str)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
b, err := d.MarshalText()
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, []byte("2021-06-08T20:12:01.000000002"), b)
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestParseDateTimeErrors(t *testing.T) {
|
func TestLocalDateTime_UnmarshalMarshalText(t *testing.T) {
|
||||||
|
d := toml.LocalDateTime{}
|
||||||
|
err := d.UnmarshalText([]byte("2021-06-08 20:12:01.000000002"))
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, toml.LocalDateTime{
|
||||||
|
toml.LocalDate{2021, 6, 8},
|
||||||
|
toml.LocalTime{20, 12, 1, 2, 9},
|
||||||
|
}, d)
|
||||||
|
|
||||||
for _, str := range []string{
|
err = d.UnmarshalText([]byte("what"))
|
||||||
"",
|
require.Error(t, err)
|
||||||
"2016-03-22", // just a date
|
|
||||||
"13:26:33", // just a time
|
err = d.UnmarshalText([]byte("2021-06-08 20:12:01.000000002 bad"))
|
||||||
"2016-03-22 13:26:33", // wrong separating character
|
require.Error(t, err)
|
||||||
"2016-03-22T13:26:33x", // extra at end
|
|
||||||
} {
|
|
||||||
if _, err := ParseLocalDateTime(str); err == nil {
|
|
||||||
t.Errorf("ParseLocalDateTime(%q) succeeded, want error", str)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestDateTimeOf(t *testing.T) {
|
|
||||||
|
|
||||||
for _, test := range []struct {
|
|
||||||
time time.Time
|
|
||||||
want LocalDateTime
|
|
||||||
}{
|
|
||||||
{
|
|
||||||
time.Date(2014, 8, 20, 15, 8, 43, 1, time.Local),
|
|
||||||
LocalDateTime{LocalDate{2014, 8, 20}, LocalTime{15, 8, 43, 1}},
|
|
||||||
},
|
|
||||||
{
|
|
||||||
time.Date(1, 1, 1, 0, 0, 0, 0, time.UTC),
|
|
||||||
LocalDateTime{LocalDate{1, 1, 1}, LocalTime{0, 0, 0, 0}},
|
|
||||||
},
|
|
||||||
} {
|
|
||||||
if got := LocalDateTimeOf(test.time); got != test.want {
|
|
||||||
t.Errorf("LocalDateTimeOf(%v) = %+v, want %+v", test.time, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestDateTimeIsValid(t *testing.T) {
|
|
||||||
|
|
||||||
// No need to be exhaustive here; it's just LocalDate.IsValid && LocalTime.IsValid.
|
|
||||||
for _, test := range []struct {
|
|
||||||
dt LocalDateTime
|
|
||||||
want bool
|
|
||||||
}{
|
|
||||||
{LocalDateTime{LocalDate{2016, 3, 20}, LocalTime{0, 0, 0, 0}}, true},
|
|
||||||
{LocalDateTime{LocalDate{2016, -3, 20}, LocalTime{0, 0, 0, 0}}, false},
|
|
||||||
{LocalDateTime{LocalDate{2016, 3, 20}, LocalTime{24, 0, 0, 0}}, false},
|
|
||||||
} {
|
|
||||||
got := test.dt.IsValid()
|
|
||||||
if got != test.want {
|
|
||||||
t.Errorf("%#v: got %t, want %t", test.dt, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestDateTimeIn(t *testing.T) {
|
|
||||||
|
|
||||||
dt := LocalDateTime{LocalDate{2016, 1, 2}, LocalTime{3, 4, 5, 6}}
|
|
||||||
|
|
||||||
want := time.Date(2016, 1, 2, 3, 4, 5, 6, time.UTC)
|
|
||||||
if got := dt.In(time.UTC); !got.Equal(want) {
|
|
||||||
t.Errorf("got %v, want %v", got, want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestDateTimeBefore(t *testing.T) {
|
|
||||||
|
|
||||||
d1 := LocalDate{2016, 12, 31}
|
|
||||||
d2 := LocalDate{2017, 1, 1}
|
|
||||||
t1 := LocalTime{5, 6, 7, 8}
|
|
||||||
t2 := LocalTime{5, 6, 7, 9}
|
|
||||||
|
|
||||||
for _, test := range []struct {
|
|
||||||
dt1, dt2 LocalDateTime
|
|
||||||
want bool
|
|
||||||
}{
|
|
||||||
{LocalDateTime{d1, t1}, LocalDateTime{d2, t1}, true},
|
|
||||||
{LocalDateTime{d1, t1}, LocalDateTime{d1, t2}, true},
|
|
||||||
{LocalDateTime{d2, t1}, LocalDateTime{d1, t1}, false},
|
|
||||||
{LocalDateTime{d2, t1}, LocalDateTime{d2, t1}, false},
|
|
||||||
} {
|
|
||||||
if got := test.dt1.Before(test.dt2); got != test.want {
|
|
||||||
t.Errorf("%v.Before(%v): got %t, want %t", test.dt1, test.dt2, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestDateTimeAfter(t *testing.T) {
|
|
||||||
|
|
||||||
d1 := LocalDate{2016, 12, 31}
|
|
||||||
d2 := LocalDate{2017, 1, 1}
|
|
||||||
t1 := LocalTime{5, 6, 7, 8}
|
|
||||||
t2 := LocalTime{5, 6, 7, 9}
|
|
||||||
|
|
||||||
for _, test := range []struct {
|
|
||||||
dt1, dt2 LocalDateTime
|
|
||||||
want bool
|
|
||||||
}{
|
|
||||||
{LocalDateTime{d1, t1}, LocalDateTime{d2, t1}, false},
|
|
||||||
{LocalDateTime{d1, t1}, LocalDateTime{d1, t2}, false},
|
|
||||||
{LocalDateTime{d2, t1}, LocalDateTime{d1, t1}, true},
|
|
||||||
{LocalDateTime{d2, t1}, LocalDateTime{d2, t1}, false},
|
|
||||||
} {
|
|
||||||
if got := test.dt1.After(test.dt2); got != test.want {
|
|
||||||
t.Errorf("%v.After(%v): got %t, want %t", test.dt1, test.dt2, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestMarshalJSON(t *testing.T) {
|
|
||||||
|
|
||||||
for _, test := range []struct {
|
|
||||||
value interface{}
|
|
||||||
want string
|
|
||||||
}{
|
|
||||||
{LocalDate{1987, 4, 15}, `"1987-04-15"`},
|
|
||||||
{LocalTime{18, 54, 2, 0}, `"18:54:02"`},
|
|
||||||
{LocalDateTime{LocalDate{1987, 4, 15}, LocalTime{18, 54, 2, 0}}, `"1987-04-15T18:54:02"`},
|
|
||||||
} {
|
|
||||||
bgot, err := json.Marshal(test.value)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if got := string(bgot); got != test.want {
|
|
||||||
t.Errorf("%#v: got %s, want %s", test.value, got, test.want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestUnmarshalJSON(t *testing.T) {
|
|
||||||
|
|
||||||
var (
|
|
||||||
d LocalDate
|
|
||||||
tm LocalTime
|
|
||||||
dt LocalDateTime
|
|
||||||
)
|
|
||||||
|
|
||||||
for _, test := range []struct {
|
|
||||||
data string
|
|
||||||
ptr interface{}
|
|
||||||
want interface{}
|
|
||||||
}{
|
|
||||||
{`"1987-04-15"`, &d, &LocalDate{1987, 4, 15}},
|
|
||||||
{`"1987-04-\u0031\u0035"`, &d, &LocalDate{1987, 4, 15}},
|
|
||||||
{`"18:54:02"`, &tm, &LocalTime{18, 54, 2, 0}},
|
|
||||||
{`"1987-04-15T18:54:02"`, &dt, &LocalDateTime{LocalDate{1987, 4, 15}, LocalTime{18, 54, 2, 0}}},
|
|
||||||
} {
|
|
||||||
if err := json.Unmarshal([]byte(test.data), test.ptr); err != nil {
|
|
||||||
t.Fatalf("%s: %v", test.data, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if !cmpEqual(test.ptr, test.want) {
|
|
||||||
t.Errorf("%s: got %#v, want %#v", test.data, test.ptr, test.want)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, bad := range []string{
|
|
||||||
"", `""`, `"bad"`, `"1987-04-15x"`,
|
|
||||||
`19870415`, // a JSON number
|
|
||||||
`11987-04-15x`, // not a JSON string
|
|
||||||
|
|
||||||
} {
|
|
||||||
if json.Unmarshal([]byte(bad), &d) == nil {
|
|
||||||
t.Errorf("%q, LocalDate: got nil, want error", bad)
|
|
||||||
}
|
|
||||||
|
|
||||||
if json.Unmarshal([]byte(bad), &tm) == nil {
|
|
||||||
t.Errorf("%q, LocalTime: got nil, want error", bad)
|
|
||||||
}
|
|
||||||
|
|
||||||
if json.Unmarshal([]byte(bad), &dt) == nil {
|
|
||||||
t.Errorf("%q, LocalDateTime: got nil, want error", bad)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|||||||
+330
-87
@@ -5,11 +5,13 @@ import (
|
|||||||
"encoding"
|
"encoding"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
|
"math"
|
||||||
"reflect"
|
"reflect"
|
||||||
"sort"
|
"sort"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
"unicode"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Marshal serializes a Go value as a TOML document.
|
// Marshal serializes a Go value as a TOML document.
|
||||||
@@ -52,9 +54,10 @@ func NewEncoder(w io.Writer) *Encoder {
|
|||||||
// This behavior can be controlled on an individual struct field basis with the
|
// This behavior can be controlled on an individual struct field basis with the
|
||||||
// inline tag:
|
// inline tag:
|
||||||
//
|
//
|
||||||
// MyField `inline:"true"`
|
// MyField `toml:",inline"`
|
||||||
func (enc *Encoder) SetTablesInline(inline bool) {
|
func (enc *Encoder) SetTablesInline(inline bool) *Encoder {
|
||||||
enc.tablesInline = inline
|
enc.tablesInline = inline
|
||||||
|
return enc
|
||||||
}
|
}
|
||||||
|
|
||||||
// SetArraysMultiline forces the encoder to emit all arrays with one element per
|
// SetArraysMultiline forces the encoder to emit all arrays with one element per
|
||||||
@@ -62,28 +65,31 @@ func (enc *Encoder) SetTablesInline(inline bool) {
|
|||||||
//
|
//
|
||||||
// This behavior can be controlled on an individual struct field basis with the multiline tag:
|
// This behavior can be controlled on an individual struct field basis with the multiline tag:
|
||||||
//
|
//
|
||||||
// MyField `multiline:"true"`
|
// MyField `multiline:"true"`
|
||||||
func (enc *Encoder) SetArraysMultiline(multiline bool) {
|
func (enc *Encoder) SetArraysMultiline(multiline bool) *Encoder {
|
||||||
enc.arraysMultiline = multiline
|
enc.arraysMultiline = multiline
|
||||||
|
return enc
|
||||||
}
|
}
|
||||||
|
|
||||||
// SetIndentSymbol defines the string that should be used for indentation. The
|
// SetIndentSymbol defines the string that should be used for indentation. The
|
||||||
// provided string is repeated for each indentation level. Defaults to two
|
// provided string is repeated for each indentation level. Defaults to two
|
||||||
// spaces.
|
// spaces.
|
||||||
func (enc *Encoder) SetIndentSymbol(s string) {
|
func (enc *Encoder) SetIndentSymbol(s string) *Encoder {
|
||||||
enc.indentSymbol = s
|
enc.indentSymbol = s
|
||||||
|
return enc
|
||||||
}
|
}
|
||||||
|
|
||||||
// SetIndentTables forces the encoder to intent tables and array tables.
|
// SetIndentTables forces the encoder to intent tables and array tables.
|
||||||
func (enc *Encoder) SetIndentTables(indent bool) {
|
func (enc *Encoder) SetIndentTables(indent bool) *Encoder {
|
||||||
enc.indentTables = indent
|
enc.indentTables = indent
|
||||||
|
return enc
|
||||||
}
|
}
|
||||||
|
|
||||||
// Encode writes a TOML representation of v to the stream.
|
// Encode writes a TOML representation of v to the stream.
|
||||||
//
|
//
|
||||||
// If v cannot be represented to TOML it returns an error.
|
// If v cannot be represented to TOML it returns an error.
|
||||||
//
|
//
|
||||||
// Encoding rules
|
// # Encoding rules
|
||||||
//
|
//
|
||||||
// A top level slice containing only maps or structs is encoded as [[table
|
// A top level slice containing only maps or structs is encoded as [[table
|
||||||
// array]].
|
// array]].
|
||||||
@@ -98,27 +104,52 @@ func (enc *Encoder) SetIndentTables(indent bool) {
|
|||||||
// Intermediate tables are always printed.
|
// Intermediate tables are always printed.
|
||||||
//
|
//
|
||||||
// By default, strings are encoded as literal string, unless they contain either
|
// By default, strings are encoded as literal string, unless they contain either
|
||||||
// a newline character or a single quote. In that case they are emitted as quoted
|
// a newline character or a single quote. In that case they are emitted as
|
||||||
// strings.
|
// quoted strings.
|
||||||
|
//
|
||||||
|
// Unsigned integers larger than math.MaxInt64 cannot be encoded. Doing so
|
||||||
|
// results in an error. This rule exists because the TOML specification only
|
||||||
|
// requires parsers to support at least the 64 bits integer range. Allowing
|
||||||
|
// larger numbers would create non-standard TOML documents, which may not be
|
||||||
|
// readable (at best) by other implementations. To encode such numbers, a
|
||||||
|
// solution is a custom type that implements encoding.TextMarshaler.
|
||||||
//
|
//
|
||||||
// When encoding structs, fields are encoded in order of definition, with their
|
// When encoding structs, fields are encoded in order of definition, with their
|
||||||
// exact name.
|
// exact name.
|
||||||
//
|
//
|
||||||
// Struct tags
|
// Tables and array tables are separated by empty lines. However, consecutive
|
||||||
|
// subtables definitions are not. For example:
|
||||||
//
|
//
|
||||||
// The following struct tags are available to tweak encoding on a per-field
|
// [top1]
|
||||||
// basis:
|
|
||||||
//
|
//
|
||||||
// toml:"foo"
|
// [top2]
|
||||||
// Changes the name of the key to use for the field to foo.
|
// [top2.child1]
|
||||||
//
|
//
|
||||||
// multiline:"true"
|
// [[array]]
|
||||||
// When the field contains a string, it will be emitted as a quoted
|
|
||||||
// multi-line TOML string.
|
|
||||||
//
|
//
|
||||||
// inline:"true"
|
// [[array]]
|
||||||
// When the field would normally be encoded as a table, it is instead
|
// [array.child2]
|
||||||
// encoded as an inline table.
|
//
|
||||||
|
// # Struct tags
|
||||||
|
//
|
||||||
|
// The encoding of each public struct field can be customized by the format
|
||||||
|
// string in the "toml" key of the struct field's tag. This follows
|
||||||
|
// encoding/json's convention. The format string starts with the name of the
|
||||||
|
// field, optionally followed by a comma-separated list of options. The name may
|
||||||
|
// be empty in order to provide options without overriding the default name.
|
||||||
|
//
|
||||||
|
// The "multiline" option emits strings as quoted multi-line TOML strings. It
|
||||||
|
// has no effect on fields that would not be encoded as strings.
|
||||||
|
//
|
||||||
|
// The "inline" option turns fields that would be emitted as tables into inline
|
||||||
|
// tables instead. It has no effect on other fields.
|
||||||
|
//
|
||||||
|
// The "omitempty" option prevents empty values or groups from being emitted.
|
||||||
|
//
|
||||||
|
// In addition to the "toml" tag struct tag, a "comment" tag can be used to emit
|
||||||
|
// a TOML comment before the value being annotated. Comments are ignored inside
|
||||||
|
// inline tables. For array tables, the comment is only present before the first
|
||||||
|
// element of the array.
|
||||||
func (enc *Encoder) Encode(v interface{}) error {
|
func (enc *Encoder) Encode(v interface{}) error {
|
||||||
var (
|
var (
|
||||||
b []byte
|
b []byte
|
||||||
@@ -146,6 +177,8 @@ func (enc *Encoder) Encode(v interface{}) error {
|
|||||||
|
|
||||||
type valueOptions struct {
|
type valueOptions struct {
|
||||||
multiline bool
|
multiline bool
|
||||||
|
omitempty bool
|
||||||
|
comment string
|
||||||
}
|
}
|
||||||
|
|
||||||
type encoderCtx struct {
|
type encoderCtx struct {
|
||||||
@@ -195,16 +228,29 @@ func (ctx *encoderCtx) isRoot() bool {
|
|||||||
return len(ctx.parentKey) == 0 && !ctx.hasKey
|
return len(ctx.parentKey) == 0 && !ctx.hasKey
|
||||||
}
|
}
|
||||||
|
|
||||||
//nolint:cyclop,funlen
|
|
||||||
func (enc *Encoder) encode(b []byte, ctx encoderCtx, v reflect.Value) ([]byte, error) {
|
func (enc *Encoder) encode(b []byte, ctx encoderCtx, v reflect.Value) ([]byte, error) {
|
||||||
if !v.IsZero() {
|
i := v.Interface()
|
||||||
i, ok := v.Interface().(time.Time)
|
|
||||||
if ok {
|
switch x := i.(type) {
|
||||||
return i.AppendFormat(b, time.RFC3339), nil
|
case time.Time:
|
||||||
|
if x.Nanosecond() > 0 {
|
||||||
|
return x.AppendFormat(b, time.RFC3339Nano), nil
|
||||||
}
|
}
|
||||||
|
return x.AppendFormat(b, time.RFC3339), nil
|
||||||
|
case LocalTime:
|
||||||
|
return append(b, x.String()...), nil
|
||||||
|
case LocalDate:
|
||||||
|
return append(b, x.String()...), nil
|
||||||
|
case LocalDateTime:
|
||||||
|
return append(b, x.String()...), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
if v.Type().Implements(textMarshalerType) {
|
hasTextMarshaler := v.Type().Implements(textMarshalerType)
|
||||||
|
if hasTextMarshaler || (v.CanAddr() && reflect.PtrTo(v.Type()).Implements(textMarshalerType)) {
|
||||||
|
if !hasTextMarshaler {
|
||||||
|
v = v.Addr()
|
||||||
|
}
|
||||||
|
|
||||||
if ctx.isRoot() {
|
if ctx.isRoot() {
|
||||||
return nil, fmt.Errorf("toml: type %s implementing the TextMarshaler interface cannot be a root element", v.Type())
|
return nil, fmt.Errorf("toml: type %s implementing the TextMarshaler interface cannot be a root element", v.Type())
|
||||||
}
|
}
|
||||||
@@ -244,9 +290,32 @@ func (enc *Encoder) encode(b []byte, ctx encoderCtx, v reflect.Value) ([]byte, e
|
|||||||
case reflect.String:
|
case reflect.String:
|
||||||
b = enc.encodeString(b, v.String(), ctx.options)
|
b = enc.encodeString(b, v.String(), ctx.options)
|
||||||
case reflect.Float32:
|
case reflect.Float32:
|
||||||
b = strconv.AppendFloat(b, v.Float(), 'f', -1, 32)
|
f := v.Float()
|
||||||
|
|
||||||
|
if math.IsNaN(f) {
|
||||||
|
b = append(b, "nan"...)
|
||||||
|
} else if f > math.MaxFloat32 {
|
||||||
|
b = append(b, "inf"...)
|
||||||
|
} else if f < -math.MaxFloat32 {
|
||||||
|
b = append(b, "-inf"...)
|
||||||
|
} else if math.Trunc(f) == f {
|
||||||
|
b = strconv.AppendFloat(b, f, 'f', 1, 32)
|
||||||
|
} else {
|
||||||
|
b = strconv.AppendFloat(b, f, 'f', -1, 32)
|
||||||
|
}
|
||||||
case reflect.Float64:
|
case reflect.Float64:
|
||||||
b = strconv.AppendFloat(b, v.Float(), 'f', -1, 64)
|
f := v.Float()
|
||||||
|
if math.IsNaN(f) {
|
||||||
|
b = append(b, "nan"...)
|
||||||
|
} else if f > math.MaxFloat64 {
|
||||||
|
b = append(b, "inf"...)
|
||||||
|
} else if f < -math.MaxFloat64 {
|
||||||
|
b = append(b, "-inf"...)
|
||||||
|
} else if math.Trunc(f) == f {
|
||||||
|
b = strconv.AppendFloat(b, f, 'f', 1, 64)
|
||||||
|
} else {
|
||||||
|
b = strconv.AppendFloat(b, f, 'f', -1, 64)
|
||||||
|
}
|
||||||
case reflect.Bool:
|
case reflect.Bool:
|
||||||
if v.Bool() {
|
if v.Bool() {
|
||||||
b = append(b, "true"...)
|
b = append(b, "true"...)
|
||||||
@@ -254,7 +323,11 @@ func (enc *Encoder) encode(b []byte, ctx encoderCtx, v reflect.Value) ([]byte, e
|
|||||||
b = append(b, "false"...)
|
b = append(b, "false"...)
|
||||||
}
|
}
|
||||||
case reflect.Uint64, reflect.Uint32, reflect.Uint16, reflect.Uint8, reflect.Uint:
|
case reflect.Uint64, reflect.Uint32, reflect.Uint16, reflect.Uint8, reflect.Uint:
|
||||||
b = strconv.AppendUint(b, v.Uint(), 10)
|
x := v.Uint()
|
||||||
|
if x > uint64(math.MaxInt64) {
|
||||||
|
return nil, fmt.Errorf("toml: not encoding uint (%d) greater than max int64 (%d)", x, int64(math.MaxInt64))
|
||||||
|
}
|
||||||
|
b = strconv.AppendUint(b, x, 10)
|
||||||
case reflect.Int64, reflect.Int32, reflect.Int16, reflect.Int8, reflect.Int:
|
case reflect.Int64, reflect.Int32, reflect.Int16, reflect.Int8, reflect.Int:
|
||||||
b = strconv.AppendInt(b, v.Int(), 10)
|
b = strconv.AppendInt(b, v.Int(), 10)
|
||||||
default:
|
default:
|
||||||
@@ -273,19 +346,19 @@ func isNil(v reflect.Value) bool {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func shouldOmitEmpty(options valueOptions, v reflect.Value) bool {
|
||||||
|
return options.omitempty && isEmptyValue(v)
|
||||||
|
}
|
||||||
|
|
||||||
func (enc *Encoder) encodeKv(b []byte, ctx encoderCtx, options valueOptions, v reflect.Value) ([]byte, error) {
|
func (enc *Encoder) encodeKv(b []byte, ctx encoderCtx, options valueOptions, v reflect.Value) ([]byte, error) {
|
||||||
var err error
|
var err error
|
||||||
|
|
||||||
if !ctx.hasKey {
|
if !ctx.inline {
|
||||||
panic("caller of encodeKv should have set the key in the context")
|
b = enc.encodeComment(ctx.indent, options.comment, b)
|
||||||
}
|
}
|
||||||
|
|
||||||
b = enc.indent(ctx.indent, b)
|
b = enc.indent(ctx.indent, b)
|
||||||
|
b = enc.encodeKey(b, ctx.key)
|
||||||
b, err = enc.encodeKey(b, ctx.key)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
b = append(b, " = "...)
|
b = append(b, " = "...)
|
||||||
|
|
||||||
// create a copy of the context because the value of a KV shouldn't
|
// create a copy of the context because the value of a KV shouldn't
|
||||||
@@ -303,6 +376,54 @@ func (enc *Encoder) encodeKv(b []byte, ctx encoderCtx, options valueOptions, v r
|
|||||||
return b, nil
|
return b, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func isEmptyValue(v reflect.Value) bool {
|
||||||
|
switch v.Kind() {
|
||||||
|
case reflect.Struct:
|
||||||
|
return isEmptyStruct(v)
|
||||||
|
case reflect.Array, reflect.Map, reflect.Slice, reflect.String:
|
||||||
|
return v.Len() == 0
|
||||||
|
case reflect.Bool:
|
||||||
|
return !v.Bool()
|
||||||
|
case reflect.Int, reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:
|
||||||
|
return v.Int() == 0
|
||||||
|
case reflect.Uint, reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64, reflect.Uintptr:
|
||||||
|
return v.Uint() == 0
|
||||||
|
case reflect.Float32, reflect.Float64:
|
||||||
|
return v.Float() == 0
|
||||||
|
case reflect.Interface, reflect.Ptr:
|
||||||
|
return v.IsNil()
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
func isEmptyStruct(v reflect.Value) bool {
|
||||||
|
// TODO: merge with walkStruct and cache.
|
||||||
|
typ := v.Type()
|
||||||
|
for i := 0; i < typ.NumField(); i++ {
|
||||||
|
fieldType := typ.Field(i)
|
||||||
|
|
||||||
|
// only consider exported fields
|
||||||
|
if fieldType.PkgPath != "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
tag := fieldType.Tag.Get("toml")
|
||||||
|
|
||||||
|
// special field name to skip field
|
||||||
|
if tag == "-" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
f := v.Field(i)
|
||||||
|
|
||||||
|
if !isEmptyValue(f) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
const literalQuote = '\''
|
const literalQuote = '\''
|
||||||
|
|
||||||
func (enc *Encoder) encodeString(b []byte, v string, options valueOptions) []byte {
|
func (enc *Encoder) encodeString(b []byte, v string, options valueOptions) []byte {
|
||||||
@@ -314,7 +435,13 @@ func (enc *Encoder) encodeString(b []byte, v string, options valueOptions) []byt
|
|||||||
}
|
}
|
||||||
|
|
||||||
func needsQuoting(v string) bool {
|
func needsQuoting(v string) bool {
|
||||||
return strings.ContainsAny(v, "'\b\f\n\r\t")
|
// TODO: vectorize
|
||||||
|
for _, b := range []byte(v) {
|
||||||
|
if b == '\'' || b == '\r' || b == '\n' || invalidAscii(b) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
// caller should have checked that the string does not contain new lines or ' .
|
// caller should have checked that the string does not contain new lines or ' .
|
||||||
@@ -326,7 +453,6 @@ func (enc *Encoder) encodeLiteralString(b []byte, v string) []byte {
|
|||||||
return b
|
return b
|
||||||
}
|
}
|
||||||
|
|
||||||
//nolint:cyclop
|
|
||||||
func (enc *Encoder) encodeQuotedString(multiline bool, b []byte, v string) []byte {
|
func (enc *Encoder) encodeQuotedString(multiline bool, b []byte, v string) []byte {
|
||||||
stringQuote := `"`
|
stringQuote := `"`
|
||||||
|
|
||||||
@@ -386,7 +512,7 @@ func (enc *Encoder) encodeQuotedString(multiline bool, b []byte, v string) []byt
|
|||||||
return b
|
return b
|
||||||
}
|
}
|
||||||
|
|
||||||
// called should have checked that the string is in A-Z / a-z / 0-9 / - / _ .
|
// caller should have checked that the string is in A-Z / a-z / 0-9 / - / _ .
|
||||||
func (enc *Encoder) encodeUnquotedKey(b []byte, v string) []byte {
|
func (enc *Encoder) encodeUnquotedKey(b []byte, v string) []byte {
|
||||||
return append(b, v...)
|
return append(b, v...)
|
||||||
}
|
}
|
||||||
@@ -396,24 +522,17 @@ func (enc *Encoder) encodeTableHeader(ctx encoderCtx, b []byte) ([]byte, error)
|
|||||||
return b, nil
|
return b, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
b = enc.encodeComment(ctx.indent, ctx.options.comment, b)
|
||||||
|
|
||||||
b = enc.indent(ctx.indent, b)
|
b = enc.indent(ctx.indent, b)
|
||||||
|
|
||||||
b = append(b, '[')
|
b = append(b, '[')
|
||||||
|
|
||||||
var err error
|
b = enc.encodeKey(b, ctx.parentKey[0])
|
||||||
|
|
||||||
b, err = enc.encodeKey(b, ctx.parentKey[0])
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, k := range ctx.parentKey[1:] {
|
for _, k := range ctx.parentKey[1:] {
|
||||||
b = append(b, '.')
|
b = append(b, '.')
|
||||||
|
b = enc.encodeKey(b, k)
|
||||||
b, err = enc.encodeKey(b, k)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
b = append(b, "]\n"...)
|
b = append(b, "]\n"...)
|
||||||
@@ -422,19 +541,19 @@ func (enc *Encoder) encodeTableHeader(ctx encoderCtx, b []byte) ([]byte, error)
|
|||||||
}
|
}
|
||||||
|
|
||||||
//nolint:cyclop
|
//nolint:cyclop
|
||||||
func (enc *Encoder) encodeKey(b []byte, k string) ([]byte, error) {
|
func (enc *Encoder) encodeKey(b []byte, k string) []byte {
|
||||||
needsQuotation := false
|
needsQuotation := false
|
||||||
cannotUseLiteral := false
|
cannotUseLiteral := false
|
||||||
|
|
||||||
|
if len(k) == 0 {
|
||||||
|
return append(b, "''"...)
|
||||||
|
}
|
||||||
|
|
||||||
for _, c := range k {
|
for _, c := range k {
|
||||||
if (c >= 'A' && c <= 'Z') || (c >= 'a' && c <= 'z') || (c >= '0' && c <= '9') || c == '-' || c == '_' {
|
if (c >= 'A' && c <= 'Z') || (c >= 'a' && c <= 'z') || (c >= '0' && c <= '9') || c == '-' || c == '_' {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
if c == '\n' {
|
|
||||||
return nil, fmt.Errorf("toml: new line characters in keys are not supported")
|
|
||||||
}
|
|
||||||
|
|
||||||
if c == literalQuote {
|
if c == literalQuote {
|
||||||
cannotUseLiteral = true
|
cannotUseLiteral = true
|
||||||
}
|
}
|
||||||
@@ -442,13 +561,17 @@ func (enc *Encoder) encodeKey(b []byte, k string) ([]byte, error) {
|
|||||||
needsQuotation = true
|
needsQuotation = true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if needsQuotation && needsQuoting(k) {
|
||||||
|
cannotUseLiteral = true
|
||||||
|
}
|
||||||
|
|
||||||
switch {
|
switch {
|
||||||
case cannotUseLiteral:
|
case cannotUseLiteral:
|
||||||
return enc.encodeQuotedString(false, b, k), nil
|
return enc.encodeQuotedString(false, b, k)
|
||||||
case needsQuotation:
|
case needsQuotation:
|
||||||
return enc.encodeLiteralString(b, k), nil
|
return enc.encodeLiteralString(b, k)
|
||||||
default:
|
default:
|
||||||
return enc.encodeUnquotedKey(b, k), nil
|
return enc.encodeUnquotedKey(b, k)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -502,18 +625,26 @@ type table struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (t *table) pushKV(k string, v reflect.Value, options valueOptions) {
|
func (t *table) pushKV(k string, v reflect.Value, options valueOptions) {
|
||||||
|
for _, e := range t.kvs {
|
||||||
|
if e.Key == k {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
t.kvs = append(t.kvs, entry{Key: k, Value: v, Options: options})
|
t.kvs = append(t.kvs, entry{Key: k, Value: v, Options: options})
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t *table) pushTable(k string, v reflect.Value, options valueOptions) {
|
func (t *table) pushTable(k string, v reflect.Value, options valueOptions) {
|
||||||
|
for _, e := range t.tables {
|
||||||
|
if e.Key == k {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
t.tables = append(t.tables, entry{Key: k, Value: v, Options: options})
|
t.tables = append(t.tables, entry{Key: k, Value: v, Options: options})
|
||||||
}
|
}
|
||||||
|
|
||||||
func (enc *Encoder) encodeStruct(b []byte, ctx encoderCtx, v reflect.Value) ([]byte, error) {
|
func walkStruct(ctx encoderCtx, t *table, v reflect.Value) {
|
||||||
var t table
|
// TODO: cache this
|
||||||
|
|
||||||
//nolint:godox
|
|
||||||
// TODO: cache this?
|
|
||||||
typ := v.Type()
|
typ := v.Type()
|
||||||
for i := 0; i < typ.NumField(); i++ {
|
for i := 0; i < typ.NumField(); i++ {
|
||||||
fieldType := typ.Field(i)
|
fieldType := typ.Field(i)
|
||||||
@@ -523,45 +654,130 @@ func (enc *Encoder) encodeStruct(b []byte, ctx encoderCtx, v reflect.Value) ([]b
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
k, ok := fieldType.Tag.Lookup("toml")
|
tag := fieldType.Tag.Get("toml")
|
||||||
if !ok {
|
|
||||||
k = fieldType.Name
|
|
||||||
}
|
|
||||||
|
|
||||||
// special field name to skip field
|
// special field name to skip field
|
||||||
if k == "-" {
|
if tag == "-" {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
|
k, opts := parseTag(tag)
|
||||||
|
if !isValidName(k) {
|
||||||
|
k = ""
|
||||||
|
}
|
||||||
|
|
||||||
f := v.Field(i)
|
f := v.Field(i)
|
||||||
|
|
||||||
|
if k == "" {
|
||||||
|
if fieldType.Anonymous {
|
||||||
|
if fieldType.Type.Kind() == reflect.Struct {
|
||||||
|
walkStruct(ctx, t, f)
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
} else {
|
||||||
|
k = fieldType.Name
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
if isNil(f) {
|
if isNil(f) {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
options := valueOptions{
|
options := valueOptions{
|
||||||
multiline: fieldBoolTag(fieldType, "multiline"),
|
multiline: opts.multiline,
|
||||||
|
omitempty: opts.omitempty,
|
||||||
|
comment: fieldType.Tag.Get("comment"),
|
||||||
}
|
}
|
||||||
|
|
||||||
inline := fieldBoolTag(fieldType, "inline")
|
if opts.inline || !willConvertToTableOrArrayTable(ctx, f) {
|
||||||
|
|
||||||
if inline || !willConvertToTableOrArrayTable(ctx, f) {
|
|
||||||
t.pushKV(k, f, options)
|
t.pushKV(k, f, options)
|
||||||
} else {
|
} else {
|
||||||
t.pushTable(k, f, options)
|
t.pushTable(k, f, options)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (enc *Encoder) encodeStruct(b []byte, ctx encoderCtx, v reflect.Value) ([]byte, error) {
|
||||||
|
var t table
|
||||||
|
|
||||||
|
walkStruct(ctx, &t, v)
|
||||||
|
|
||||||
return enc.encodeTable(b, ctx, t)
|
return enc.encodeTable(b, ctx, t)
|
||||||
}
|
}
|
||||||
|
|
||||||
func fieldBoolTag(field reflect.StructField, tag string) bool {
|
func (enc *Encoder) encodeComment(indent int, comment string, b []byte) []byte {
|
||||||
x, ok := field.Tag.Lookup(tag)
|
for len(comment) > 0 {
|
||||||
|
var line string
|
||||||
return ok && x == "true"
|
idx := strings.IndexByte(comment, '\n')
|
||||||
|
if idx >= 0 {
|
||||||
|
line = comment[:idx]
|
||||||
|
comment = comment[idx+1:]
|
||||||
|
} else {
|
||||||
|
line = comment
|
||||||
|
comment = ""
|
||||||
|
}
|
||||||
|
b = enc.indent(indent, b)
|
||||||
|
b = append(b, "# "...)
|
||||||
|
b = append(b, line...)
|
||||||
|
b = append(b, '\n')
|
||||||
|
}
|
||||||
|
return b
|
||||||
|
}
|
||||||
|
|
||||||
|
func isValidName(s string) bool {
|
||||||
|
if s == "" {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
for _, c := range s {
|
||||||
|
switch {
|
||||||
|
case strings.ContainsRune("!#$%&()*+-./:;<=>?@[]^_{|}~ ", c):
|
||||||
|
// Backslash and quote chars are reserved, but
|
||||||
|
// otherwise any punctuation chars are allowed
|
||||||
|
// in a tag name.
|
||||||
|
case !unicode.IsLetter(c) && !unicode.IsDigit(c):
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
type tagOptions struct {
|
||||||
|
multiline bool
|
||||||
|
inline bool
|
||||||
|
omitempty bool
|
||||||
|
}
|
||||||
|
|
||||||
|
func parseTag(tag string) (string, tagOptions) {
|
||||||
|
opts := tagOptions{}
|
||||||
|
|
||||||
|
idx := strings.Index(tag, ",")
|
||||||
|
if idx == -1 {
|
||||||
|
return tag, opts
|
||||||
|
}
|
||||||
|
|
||||||
|
raw := tag[idx+1:]
|
||||||
|
tag = string(tag[:idx])
|
||||||
|
for raw != "" {
|
||||||
|
var o string
|
||||||
|
i := strings.Index(raw, ",")
|
||||||
|
if i >= 0 {
|
||||||
|
o, raw = raw[:i], raw[i+1:]
|
||||||
|
} else {
|
||||||
|
o, raw = raw, ""
|
||||||
|
}
|
||||||
|
switch o {
|
||||||
|
case "multiline":
|
||||||
|
opts.multiline = true
|
||||||
|
case "inline":
|
||||||
|
opts.inline = true
|
||||||
|
case "omitempty":
|
||||||
|
opts.omitempty = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return tag, opts
|
||||||
}
|
}
|
||||||
|
|
||||||
//nolint:cyclop
|
|
||||||
func (enc *Encoder) encodeTable(b []byte, ctx encoderCtx, t table) ([]byte, error) {
|
func (enc *Encoder) encodeTable(b []byte, ctx encoderCtx, t table) ([]byte, error) {
|
||||||
var err error
|
var err error
|
||||||
|
|
||||||
@@ -583,7 +799,13 @@ func (enc *Encoder) encodeTable(b []byte, ctx encoderCtx, t table) ([]byte, erro
|
|||||||
}
|
}
|
||||||
ctx.skipTableHeader = false
|
ctx.skipTableHeader = false
|
||||||
|
|
||||||
|
hasNonEmptyKV := false
|
||||||
for _, kv := range t.kvs {
|
for _, kv := range t.kvs {
|
||||||
|
if shouldOmitEmpty(kv.Options, kv.Value) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
hasNonEmptyKV = true
|
||||||
|
|
||||||
ctx.setKey(kv.Key)
|
ctx.setKey(kv.Key)
|
||||||
|
|
||||||
b, err = enc.encodeKv(b, ctx, kv.Options, kv.Value)
|
b, err = enc.encodeKv(b, ctx, kv.Options, kv.Value)
|
||||||
@@ -594,7 +816,20 @@ func (enc *Encoder) encodeTable(b []byte, ctx encoderCtx, t table) ([]byte, erro
|
|||||||
b = append(b, '\n')
|
b = append(b, '\n')
|
||||||
}
|
}
|
||||||
|
|
||||||
|
first := true
|
||||||
for _, table := range t.tables {
|
for _, table := range t.tables {
|
||||||
|
if shouldOmitEmpty(table.Options, table.Value) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if first {
|
||||||
|
first = false
|
||||||
|
if hasNonEmptyKV {
|
||||||
|
b = append(b, '\n')
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
b = append(b, "\n"...)
|
||||||
|
}
|
||||||
|
|
||||||
ctx.setKey(table.Key)
|
ctx.setKey(table.Key)
|
||||||
|
|
||||||
ctx.options = table.Options
|
ctx.options = table.Options
|
||||||
@@ -603,8 +838,6 @@ func (enc *Encoder) encodeTable(b []byte, ctx encoderCtx, t table) ([]byte, erro
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
b = append(b, '\n')
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return b, nil
|
return b, nil
|
||||||
@@ -617,6 +850,10 @@ func (enc *Encoder) encodeTableInline(b []byte, ctx encoderCtx, t table) ([]byte
|
|||||||
|
|
||||||
first := true
|
first := true
|
||||||
for _, kv := range t.kvs {
|
for _, kv := range t.kvs {
|
||||||
|
if shouldOmitEmpty(kv.Options, kv.Value) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
if first {
|
if first {
|
||||||
first = false
|
first = false
|
||||||
} else {
|
} else {
|
||||||
@@ -632,7 +869,7 @@ func (enc *Encoder) encodeTableInline(b []byte, ctx encoderCtx, t table) ([]byte
|
|||||||
}
|
}
|
||||||
|
|
||||||
if len(t.tables) > 0 {
|
if len(t.tables) > 0 {
|
||||||
panic("inline table cannot contain nested tables, online key-values")
|
panic("inline table cannot contain nested tables, only key-values")
|
||||||
}
|
}
|
||||||
|
|
||||||
b = append(b, "}"...)
|
b = append(b, "}"...)
|
||||||
@@ -644,7 +881,7 @@ func willConvertToTable(ctx encoderCtx, v reflect.Value) bool {
|
|||||||
if !v.IsValid() {
|
if !v.IsValid() {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
if v.Type() == timeType || v.Type().Implements(textMarshalerType) {
|
if v.Type() == timeType || v.Type().Implements(textMarshalerType) || (v.Kind() != reflect.Ptr && v.CanAddr() && reflect.PtrTo(v.Type()).Implements(textMarshalerType)) {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -666,6 +903,9 @@ func willConvertToTable(ctx encoderCtx, v reflect.Value) bool {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func willConvertToTableOrArrayTable(ctx encoderCtx, v reflect.Value) bool {
|
func willConvertToTableOrArrayTable(ctx encoderCtx, v reflect.Value) bool {
|
||||||
|
if ctx.insideKv {
|
||||||
|
return false
|
||||||
|
}
|
||||||
t := v.Type()
|
t := v.Type()
|
||||||
|
|
||||||
if t.Kind() == reflect.Interface {
|
if t.Kind() == reflect.Interface {
|
||||||
@@ -711,7 +951,6 @@ func (enc *Encoder) encodeSlice(b []byte, ctx encoderCtx, v reflect.Value) ([]by
|
|||||||
func (enc *Encoder) encodeSliceAsArrayTable(b []byte, ctx encoderCtx, v reflect.Value) ([]byte, error) {
|
func (enc *Encoder) encodeSliceAsArrayTable(b []byte, ctx encoderCtx, v reflect.Value) ([]byte, error) {
|
||||||
ctx.shiftKey()
|
ctx.shiftKey()
|
||||||
|
|
||||||
var err error
|
|
||||||
scratch := make([]byte, 0, 64)
|
scratch := make([]byte, 0, 64)
|
||||||
scratch = append(scratch, "[["...)
|
scratch = append(scratch, "[["...)
|
||||||
|
|
||||||
@@ -720,18 +959,22 @@ func (enc *Encoder) encodeSliceAsArrayTable(b []byte, ctx encoderCtx, v reflect.
|
|||||||
scratch = append(scratch, '.')
|
scratch = append(scratch, '.')
|
||||||
}
|
}
|
||||||
|
|
||||||
scratch, err = enc.encodeKey(scratch, k)
|
scratch = enc.encodeKey(scratch, k)
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
scratch = append(scratch, "]]\n"...)
|
scratch = append(scratch, "]]\n"...)
|
||||||
ctx.skipTableHeader = true
|
ctx.skipTableHeader = true
|
||||||
|
|
||||||
|
b = enc.encodeComment(ctx.indent, ctx.options.comment, b)
|
||||||
|
|
||||||
for i := 0; i < v.Len(); i++ {
|
for i := 0; i < v.Len(); i++ {
|
||||||
|
if i != 0 {
|
||||||
|
b = append(b, "\n"...)
|
||||||
|
}
|
||||||
|
|
||||||
b = append(b, scratch...)
|
b = append(b, scratch...)
|
||||||
|
|
||||||
|
var err error
|
||||||
b, err = enc.encode(b, ctx, v.Index(i))
|
b, err = enc.encode(b, ctx, v.Index(i))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
|
|||||||
+491
-83
@@ -4,20 +4,28 @@ import (
|
|||||||
"bytes"
|
"bytes"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"math"
|
||||||
|
"math/big"
|
||||||
"strings"
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/pelletier/go-toml/v2"
|
"github.com/pelletier/go-toml/v2"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
//nolint:funlen
|
|
||||||
func TestMarshal(t *testing.T) {
|
func TestMarshal(t *testing.T) {
|
||||||
someInt := 42
|
someInt := 42
|
||||||
|
|
||||||
type structInline struct {
|
type structInline struct {
|
||||||
A interface{} `inline:"true"`
|
A interface{} `toml:",inline"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type comments struct {
|
||||||
|
One int
|
||||||
|
Two int `comment:"Before kv"`
|
||||||
|
Three []int `comment:"Before array"`
|
||||||
}
|
}
|
||||||
|
|
||||||
examples := []struct {
|
examples := []struct {
|
||||||
@@ -31,21 +39,21 @@ func TestMarshal(t *testing.T) {
|
|||||||
v: map[string]string{
|
v: map[string]string{
|
||||||
"hello": "world",
|
"hello": "world",
|
||||||
},
|
},
|
||||||
expected: "hello = 'world'",
|
expected: "hello = 'world'\n",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "map with new line in key",
|
desc: "map with new line in key",
|
||||||
v: map[string]string{
|
v: map[string]string{
|
||||||
"hel\nlo": "world",
|
"hel\nlo": "world",
|
||||||
},
|
},
|
||||||
err: true,
|
expected: "\"hel\\nlo\" = 'world'\n",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: `map with " in key`,
|
desc: `map with " in key`,
|
||||||
v: map[string]string{
|
v: map[string]string{
|
||||||
`hel"lo`: "world",
|
`hel"lo`: "world",
|
||||||
},
|
},
|
||||||
expected: `'hel"lo' = 'world'`,
|
expected: "'hel\"lo' = 'world'\n",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "map in map and string",
|
desc: "map in map and string",
|
||||||
@@ -54,9 +62,9 @@ func TestMarshal(t *testing.T) {
|
|||||||
"hello": "world",
|
"hello": "world",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `[table]
|
||||||
[table]
|
hello = 'world'
|
||||||
hello = 'world'`,
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "map in map in map and string",
|
desc: "map in map in map and string",
|
||||||
@@ -67,10 +75,10 @@ hello = 'world'`,
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `[this]
|
||||||
[this]
|
|
||||||
[this.is]
|
[this.is]
|
||||||
a = 'test'`,
|
a = 'test'
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "map in map in map and string with values",
|
desc: "map in map in map and string with values",
|
||||||
@@ -82,18 +90,20 @@ a = 'test'`,
|
|||||||
"also": "that",
|
"also": "that",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `[this]
|
||||||
[this]
|
|
||||||
also = 'that'
|
also = 'that'
|
||||||
|
|
||||||
[this.is]
|
[this.is]
|
||||||
a = 'test'`,
|
a = 'test'
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "simple string array",
|
desc: "simple string array",
|
||||||
v: map[string][]string{
|
v: map[string][]string{
|
||||||
"array": {"one", "two", "three"},
|
"array": {"one", "two", "three"},
|
||||||
},
|
},
|
||||||
expected: `array = ['one', 'two', 'three']`,
|
expected: `array = ['one', 'two', 'three']
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "empty string array",
|
desc: "empty string array",
|
||||||
@@ -110,14 +120,16 @@ a = 'test'`,
|
|||||||
v: map[string][][]string{
|
v: map[string][][]string{
|
||||||
"array": {{"one", "two"}, {"three"}},
|
"array": {{"one", "two"}, {"three"}},
|
||||||
},
|
},
|
||||||
expected: `array = [['one', 'two'], ['three']]`,
|
expected: `array = [['one', 'two'], ['three']]
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "mixed strings and nested string arrays",
|
desc: "mixed strings and nested string arrays",
|
||||||
v: map[string][]interface{}{
|
v: map[string][]interface{}{
|
||||||
"array": {"a string", []string{"one", "two"}, "last"},
|
"array": {"a string", []string{"one", "two"}, "last"},
|
||||||
},
|
},
|
||||||
expected: `array = ['a string', ['one', 'two'], 'last']`,
|
expected: `array = ['a string', ['one', 'two'], 'last']
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "array of maps",
|
desc: "array of maps",
|
||||||
@@ -127,9 +139,9 @@ a = 'test'`,
|
|||||||
{"map2.1": "v2.1"},
|
{"map2.1": "v2.1"},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `[[top]]
|
||||||
[[top]]
|
|
||||||
'map1.1' = 'v1.1'
|
'map1.1' = 'v1.1'
|
||||||
|
|
||||||
[[top]]
|
[[top]]
|
||||||
'map2.1' = 'v2.1'
|
'map2.1' = 'v2.1'
|
||||||
`,
|
`,
|
||||||
@@ -140,9 +152,9 @@ a = 'test'`,
|
|||||||
"key1": "value1",
|
"key1": "value1",
|
||||||
"key2": "value2",
|
"key2": "value2",
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `key1 = 'value1'
|
||||||
key1 = 'value1'
|
key2 = 'value2'
|
||||||
key2 = 'value2'`,
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "simple struct",
|
desc: "simple struct",
|
||||||
@@ -151,7 +163,8 @@ key2 = 'value2'`,
|
|||||||
}{
|
}{
|
||||||
A: "foo",
|
A: "foo",
|
||||||
},
|
},
|
||||||
expected: `A = 'foo'`,
|
expected: `A = 'foo'
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "one level of structs within structs",
|
desc: "one level of structs within structs",
|
||||||
@@ -166,8 +179,7 @@ key2 = 'value2'`,
|
|||||||
K2: "v2",
|
K2: "v2",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `[A]
|
||||||
[A]
|
|
||||||
K1 = 'v1'
|
K1 = 'v1'
|
||||||
K2 = 'v2'
|
K2 = 'v2'
|
||||||
`,
|
`,
|
||||||
@@ -182,10 +194,10 @@ K2 = 'v2'
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `[root]
|
||||||
[root]
|
|
||||||
[[root.nested]]
|
[[root.nested]]
|
||||||
name = 'Bob'
|
name = 'Bob'
|
||||||
|
|
||||||
[[root.nested]]
|
[[root.nested]]
|
||||||
name = 'Alice'
|
name = 'Alice'
|
||||||
`,
|
`,
|
||||||
@@ -193,67 +205,72 @@ name = 'Alice'
|
|||||||
{
|
{
|
||||||
desc: "string escapes",
|
desc: "string escapes",
|
||||||
v: map[string]interface{}{
|
v: map[string]interface{}{
|
||||||
"a": `'"\`,
|
"a": "'\b\f\r\t\"\\",
|
||||||
},
|
},
|
||||||
expected: `a = "'\"\\"`,
|
expected: `a = "'\b\f\r\t\"\\"
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "string utf8 low",
|
desc: "string utf8 low",
|
||||||
v: map[string]interface{}{
|
v: map[string]interface{}{
|
||||||
"a": "'Ę",
|
"a": "'Ę",
|
||||||
},
|
},
|
||||||
expected: `a = "'Ę"`,
|
expected: `a = "'Ę"
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "string utf8 low 2",
|
desc: "string utf8 low 2",
|
||||||
v: map[string]interface{}{
|
v: map[string]interface{}{
|
||||||
"a": "'\u10A85",
|
"a": "'\u10A85",
|
||||||
},
|
},
|
||||||
expected: "a = \"'\u10A85\"",
|
expected: "a = \"'\u10A85\"\n",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "string utf8 low 2",
|
desc: "string utf8 low 2",
|
||||||
v: map[string]interface{}{
|
v: map[string]interface{}{
|
||||||
"a": "'\u10A85",
|
"a": "'\u10A85",
|
||||||
},
|
},
|
||||||
expected: "a = \"'\u10A85\"",
|
expected: "a = \"'\u10A85\"\n",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "emoji",
|
desc: "emoji",
|
||||||
v: map[string]interface{}{
|
v: map[string]interface{}{
|
||||||
"a": "'😀",
|
"a": "'😀",
|
||||||
},
|
},
|
||||||
expected: "a = \"'😀\"",
|
expected: "a = \"'😀\"\n",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "control char",
|
desc: "control char",
|
||||||
v: map[string]interface{}{
|
v: map[string]interface{}{
|
||||||
"a": "'\u001A",
|
"a": "'\u001A",
|
||||||
},
|
},
|
||||||
expected: `a = "'\u001A"`,
|
expected: `a = "'\u001A"
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "multi-line string",
|
desc: "multi-line string",
|
||||||
v: map[string]interface{}{
|
v: map[string]interface{}{
|
||||||
"a": "hello\nworld",
|
"a": "hello\nworld",
|
||||||
},
|
},
|
||||||
expected: `a = "hello\nworld"`,
|
expected: `a = "hello\nworld"
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "multi-line forced",
|
desc: "multi-line forced",
|
||||||
v: struct {
|
v: struct {
|
||||||
A string `multiline:"true"`
|
A string `toml:",multiline"`
|
||||||
}{
|
}{
|
||||||
A: "hello\nworld",
|
A: "hello\nworld",
|
||||||
},
|
},
|
||||||
expected: `A = """
|
expected: `A = """
|
||||||
hello
|
hello
|
||||||
world"""`,
|
world"""
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "inline field",
|
desc: "inline field",
|
||||||
v: struct {
|
v: struct {
|
||||||
A map[string]string `inline:"true"`
|
A map[string]string `toml:",inline"`
|
||||||
B map[string]string
|
B map[string]string
|
||||||
}{
|
}{
|
||||||
A: map[string]string{
|
A: map[string]string{
|
||||||
@@ -263,8 +280,8 @@ world"""`,
|
|||||||
"isinline": "no",
|
"isinline": "no",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `A = {isinline = 'yes'}
|
||||||
A = {isinline = 'yes'}
|
|
||||||
[B]
|
[B]
|
||||||
isinline = 'no'
|
isinline = 'no'
|
||||||
`,
|
`,
|
||||||
@@ -272,14 +289,13 @@ isinline = 'no'
|
|||||||
{
|
{
|
||||||
desc: "mutiline array int",
|
desc: "mutiline array int",
|
||||||
v: struct {
|
v: struct {
|
||||||
A []int `multiline:"true"`
|
A []int `toml:",multiline"`
|
||||||
B []int
|
B []int
|
||||||
}{
|
}{
|
||||||
A: []int{1, 2, 3, 4},
|
A: []int{1, 2, 3, 4},
|
||||||
B: []int{1, 2, 3, 4},
|
B: []int{1, 2, 3, 4},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `A = [
|
||||||
A = [
|
|
||||||
1,
|
1,
|
||||||
2,
|
2,
|
||||||
3,
|
3,
|
||||||
@@ -291,12 +307,11 @@ B = [1, 2, 3, 4]
|
|||||||
{
|
{
|
||||||
desc: "mutiline array in array",
|
desc: "mutiline array in array",
|
||||||
v: struct {
|
v: struct {
|
||||||
A [][]int `multiline:"true"`
|
A [][]int `toml:",multiline"`
|
||||||
}{
|
}{
|
||||||
A: [][]int{{1, 2}, {3, 4}},
|
A: [][]int{{1, 2}, {3, 4}},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `A = [
|
||||||
A = [
|
|
||||||
[1, 2],
|
[1, 2],
|
||||||
[3, 4]
|
[3, 4]
|
||||||
]
|
]
|
||||||
@@ -321,7 +336,8 @@ A = [
|
|||||||
}{
|
}{
|
||||||
A: []*int{nil},
|
A: []*int{nil},
|
||||||
},
|
},
|
||||||
expected: `A = [0]`,
|
expected: `A = [0]
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "nil pointer in slice uses zero value",
|
desc: "nil pointer in slice uses zero value",
|
||||||
@@ -330,7 +346,8 @@ A = [
|
|||||||
}{
|
}{
|
||||||
A: []*int{nil},
|
A: []*int{nil},
|
||||||
},
|
},
|
||||||
expected: `A = [0]`,
|
expected: `A = [0]
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "pointer in slice",
|
desc: "pointer in slice",
|
||||||
@@ -339,7 +356,8 @@ A = [
|
|||||||
}{
|
}{
|
||||||
A: []*int{&someInt},
|
A: []*int{&someInt},
|
||||||
},
|
},
|
||||||
expected: `A = [42]`,
|
expected: `A = [42]
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "inline table in inline table",
|
desc: "inline table in inline table",
|
||||||
@@ -350,30 +368,34 @@ A = [
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
expected: `A = {A = {A = 'hello'}}`,
|
expected: `A = {A = {A = 'hello'}}
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "empty slice in map",
|
desc: "empty slice in map",
|
||||||
v: map[string][]string{
|
v: map[string][]string{
|
||||||
"a": {},
|
"a": {},
|
||||||
},
|
},
|
||||||
expected: `a = []`,
|
expected: `a = []
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "map in slice",
|
desc: "map in slice",
|
||||||
v: map[string][]map[string]string{
|
v: map[string][]map[string]string{
|
||||||
"a": {{"hello": "world"}},
|
"a": {{"hello": "world"}},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `[[a]]
|
||||||
[[a]]
|
hello = 'world'
|
||||||
hello = 'world'`,
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "newline in map in slice",
|
desc: "newline in map in slice",
|
||||||
v: map[string][]map[string]string{
|
v: map[string][]map[string]string{
|
||||||
"a\n": {{"hello": "world"}},
|
"a\n": {{"hello": "world"}},
|
||||||
},
|
},
|
||||||
err: true,
|
expected: `[["a\n"]]
|
||||||
|
hello = 'world'
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "newline in map in slice",
|
desc: "newline in map in slice",
|
||||||
@@ -389,7 +411,8 @@ hello = 'world'`,
|
|||||||
}{
|
}{
|
||||||
A: []struct{}{},
|
A: []struct{}{},
|
||||||
},
|
},
|
||||||
expected: `A = []`,
|
expected: `A = []
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "nil field is ignored",
|
desc: "nil field is ignored",
|
||||||
@@ -409,7 +432,8 @@ hello = 'world'`,
|
|||||||
Public: "shown",
|
Public: "shown",
|
||||||
private: "hidden",
|
private: "hidden",
|
||||||
},
|
},
|
||||||
expected: `Public = 'shown'`,
|
expected: `Public = 'shown'
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "fields tagged - are ignored",
|
desc: "fields tagged - are ignored",
|
||||||
@@ -433,7 +457,8 @@ hello = 'world'`,
|
|||||||
v: map[string]interface{}{
|
v: map[string]interface{}{
|
||||||
"hello\nworld": 42,
|
"hello\nworld": 42,
|
||||||
},
|
},
|
||||||
err: true,
|
expected: `"hello\nworld" = 42
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "new line in parent of nested table key",
|
desc: "new line in parent of nested table key",
|
||||||
@@ -442,7 +467,9 @@ hello = 'world'`,
|
|||||||
"inner": 42,
|
"inner": 42,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
err: true,
|
expected: `["hello\nworld"]
|
||||||
|
inner = 42
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "new line in nested table key",
|
desc: "new line in nested table key",
|
||||||
@@ -453,7 +480,10 @@ hello = 'world'`,
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
err: true,
|
expected: `[parent]
|
||||||
|
[parent."in\ner"]
|
||||||
|
foo = 42
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "invalid map key",
|
desc: "invalid map key",
|
||||||
@@ -469,6 +499,39 @@ hello = 'world'`,
|
|||||||
},
|
},
|
||||||
err: true,
|
err: true,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
desc: "time",
|
||||||
|
v: struct {
|
||||||
|
T time.Time
|
||||||
|
}{
|
||||||
|
T: time.Time{},
|
||||||
|
},
|
||||||
|
expected: `T = 0001-01-01T00:00:00Z
|
||||||
|
`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
desc: "time nano",
|
||||||
|
v: struct {
|
||||||
|
T time.Time
|
||||||
|
}{
|
||||||
|
T: time.Date(1979, time.May, 27, 0, 32, 0, 999999000, time.UTC),
|
||||||
|
},
|
||||||
|
expected: `T = 1979-05-27T00:32:00.999999Z
|
||||||
|
`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
desc: "bool",
|
||||||
|
v: struct {
|
||||||
|
A bool
|
||||||
|
B bool
|
||||||
|
}{
|
||||||
|
A: false,
|
||||||
|
B: true,
|
||||||
|
},
|
||||||
|
expected: `A = false
|
||||||
|
B = true
|
||||||
|
`,
|
||||||
|
},
|
||||||
{
|
{
|
||||||
desc: "numbers",
|
desc: "numbers",
|
||||||
v: struct {
|
v: struct {
|
||||||
@@ -483,6 +546,7 @@ hello = 'world'`,
|
|||||||
I int16
|
I int16
|
||||||
J int8
|
J int8
|
||||||
K int
|
K int
|
||||||
|
L float64
|
||||||
}{
|
}{
|
||||||
A: 1.1,
|
A: 1.1,
|
||||||
B: 42,
|
B: 42,
|
||||||
@@ -495,9 +559,9 @@ hello = 'world'`,
|
|||||||
I: 42,
|
I: 42,
|
||||||
J: 42,
|
J: 42,
|
||||||
K: 42,
|
K: 42,
|
||||||
|
L: 2.2,
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `A = 1.1
|
||||||
A = 1.1
|
|
||||||
B = 42
|
B = 42
|
||||||
C = 42
|
C = 42
|
||||||
D = 42
|
D = 42
|
||||||
@@ -507,7 +571,29 @@ G = 42
|
|||||||
H = 42
|
H = 42
|
||||||
I = 42
|
I = 42
|
||||||
J = 42
|
J = 42
|
||||||
K = 42`,
|
K = 42
|
||||||
|
L = 2.2
|
||||||
|
`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
desc: "comments",
|
||||||
|
v: struct {
|
||||||
|
Table comments `comment:"Before table"`
|
||||||
|
}{
|
||||||
|
Table: comments{
|
||||||
|
One: 1,
|
||||||
|
Two: 2,
|
||||||
|
Three: []int{1, 2, 3},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
expected: `# Before table
|
||||||
|
[Table]
|
||||||
|
One = 1
|
||||||
|
# Before kv
|
||||||
|
Two = 2
|
||||||
|
# Before array
|
||||||
|
Three = [1, 2, 3]
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -522,7 +608,7 @@ K = 42`,
|
|||||||
}
|
}
|
||||||
|
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
equalStringsIgnoreNewlines(t, e.expected, string(b))
|
assert.Equal(t, e.expected, string(b))
|
||||||
|
|
||||||
// make sure the output is always valid TOML
|
// make sure the output is always valid TOML
|
||||||
defaultMap := map[string]interface{}{}
|
defaultMap := map[string]interface{}{}
|
||||||
@@ -551,7 +637,7 @@ K = 42`,
|
|||||||
|
|
||||||
type flagsSetters []struct {
|
type flagsSetters []struct {
|
||||||
name string
|
name string
|
||||||
f func(enc *toml.Encoder, flag bool)
|
f func(enc *toml.Encoder, flag bool) *toml.Encoder
|
||||||
}
|
}
|
||||||
|
|
||||||
var allFlags = flagsSetters{
|
var allFlags = flagsSetters{
|
||||||
@@ -597,10 +683,31 @@ func testWithFlags(t *testing.T, flags int, setters flagsSetters, testfn func(t
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func equalStringsIgnoreNewlines(t *testing.T, expected string, actual string) {
|
func TestMarshalFloats(t *testing.T) {
|
||||||
t.Helper()
|
v := map[string]float32{
|
||||||
cutset := "\n"
|
"nan": float32(math.NaN()),
|
||||||
assert.Equal(t, strings.Trim(expected, cutset), strings.Trim(actual, cutset))
|
"+inf": float32(math.Inf(1)),
|
||||||
|
"-inf": float32(math.Inf(-1)),
|
||||||
|
}
|
||||||
|
|
||||||
|
expected := `'+inf' = inf
|
||||||
|
-inf = -inf
|
||||||
|
nan = nan
|
||||||
|
`
|
||||||
|
|
||||||
|
actual, err := toml.Marshal(v)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, expected, string(actual))
|
||||||
|
|
||||||
|
v64 := map[string]float64{
|
||||||
|
"nan": math.NaN(),
|
||||||
|
"+inf": math.Inf(1),
|
||||||
|
"-inf": math.Inf(-1),
|
||||||
|
}
|
||||||
|
|
||||||
|
actual, err = toml.Marshal(v64)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, expected, string(actual))
|
||||||
}
|
}
|
||||||
|
|
||||||
//nolint:funlen
|
//nolint:funlen
|
||||||
@@ -615,7 +722,8 @@ func TestMarshalIndentTables(t *testing.T) {
|
|||||||
v: map[string]interface{}{
|
v: map[string]interface{}{
|
||||||
"foo": "bar",
|
"foo": "bar",
|
||||||
},
|
},
|
||||||
expected: `foo = 'bar'`,
|
expected: `foo = 'bar'
|
||||||
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
desc: "one level table",
|
desc: "one level table",
|
||||||
@@ -625,8 +733,7 @@ func TestMarshalIndentTables(t *testing.T) {
|
|||||||
"two": "value2",
|
"two": "value2",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `[foo]
|
||||||
[foo]
|
|
||||||
one = 'value1'
|
one = 'value1'
|
||||||
two = 'value2'
|
two = 'value2'
|
||||||
`,
|
`,
|
||||||
@@ -642,10 +749,11 @@ func TestMarshalIndentTables(t *testing.T) {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
expected: `
|
expected: `root = 'value0'
|
||||||
root = 'value0'
|
|
||||||
[level1]
|
[level1]
|
||||||
one = 'value1'
|
one = 'value1'
|
||||||
|
|
||||||
[level1.level2]
|
[level1.level2]
|
||||||
two = 'value2'
|
two = 'value2'
|
||||||
`,
|
`,
|
||||||
@@ -660,7 +768,7 @@ root = 'value0'
|
|||||||
enc.SetIndentTables(true)
|
enc.SetIndentTables(true)
|
||||||
err := enc.Encode(e.v)
|
err := enc.Encode(e.v)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
equalStringsIgnoreNewlines(t, e.expected, buf.String())
|
assert.Equal(t, e.expected, buf.String())
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -705,7 +813,7 @@ func TestMarshalTextMarshaler(t *testing.T) {
|
|||||||
m := map[string]interface{}{"a": &customTextMarshaler{value: 2}}
|
m := map[string]interface{}{"a": &customTextMarshaler{value: 2}}
|
||||||
r, err := toml.Marshal(m)
|
r, err := toml.Marshal(m)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
equalStringsIgnoreNewlines(t, "a = '::2'", string(r))
|
assert.Equal(t, "a = '::2'\n", string(r))
|
||||||
}
|
}
|
||||||
|
|
||||||
type brokenWriter struct{}
|
type brokenWriter struct{}
|
||||||
@@ -728,10 +836,63 @@ func TestEncoderSetIndentSymbol(t *testing.T) {
|
|||||||
enc.SetIndentSymbol(">>>")
|
enc.SetIndentSymbol(">>>")
|
||||||
err := enc.Encode(map[string]map[string]string{"parent": {"hello": "world"}})
|
err := enc.Encode(map[string]map[string]string{"parent": {"hello": "world"}})
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
expected := `
|
expected := `[parent]
|
||||||
[parent]
|
>>>hello = 'world'
|
||||||
>>>hello = 'world'`
|
`
|
||||||
equalStringsIgnoreNewlines(t, expected, w.String())
|
assert.Equal(t, expected, w.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestEncoderOmitempty(t *testing.T) {
|
||||||
|
type doc struct {
|
||||||
|
String string `toml:",omitempty,multiline"`
|
||||||
|
Bool bool `toml:",omitempty,multiline"`
|
||||||
|
Int int `toml:",omitempty,multiline"`
|
||||||
|
Int8 int8 `toml:",omitempty,multiline"`
|
||||||
|
Int16 int16 `toml:",omitempty,multiline"`
|
||||||
|
Int32 int32 `toml:",omitempty,multiline"`
|
||||||
|
Int64 int64 `toml:",omitempty,multiline"`
|
||||||
|
Uint uint `toml:",omitempty,multiline"`
|
||||||
|
Uint8 uint8 `toml:",omitempty,multiline"`
|
||||||
|
Uint16 uint16 `toml:",omitempty,multiline"`
|
||||||
|
Uint32 uint32 `toml:",omitempty,multiline"`
|
||||||
|
Uint64 uint64 `toml:",omitempty,multiline"`
|
||||||
|
Float32 float32 `toml:",omitempty,multiline"`
|
||||||
|
Float64 float64 `toml:",omitempty,multiline"`
|
||||||
|
MapNil map[string]string `toml:",omitempty,multiline"`
|
||||||
|
Slice []string `toml:",omitempty,multiline"`
|
||||||
|
Ptr *string `toml:",omitempty,multiline"`
|
||||||
|
Iface interface{} `toml:",omitempty,multiline"`
|
||||||
|
Struct struct{} `toml:",omitempty,multiline"`
|
||||||
|
}
|
||||||
|
|
||||||
|
d := doc{}
|
||||||
|
|
||||||
|
b, err := toml.Marshal(d)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
expected := ``
|
||||||
|
|
||||||
|
assert.Equal(t, expected, string(b))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestEncoderTagFieldName(t *testing.T) {
|
||||||
|
type doc struct {
|
||||||
|
String string `toml:"hello"`
|
||||||
|
OkSym string `toml:"#"`
|
||||||
|
Bad string `toml:"\"`
|
||||||
|
}
|
||||||
|
|
||||||
|
d := doc{String: "world"}
|
||||||
|
|
||||||
|
b, err := toml.Marshal(d)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
expected := `hello = 'world'
|
||||||
|
'#' = ''
|
||||||
|
Bad = ''
|
||||||
|
`
|
||||||
|
|
||||||
|
assert.Equal(t, expected, string(b))
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestIssue436(t *testing.T) {
|
func TestIssue436(t *testing.T) {
|
||||||
@@ -745,12 +906,11 @@ func TestIssue436(t *testing.T) {
|
|||||||
err = toml.NewEncoder(&buf).Encode(v)
|
err = toml.NewEncoder(&buf).Encode(v)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
expected := `
|
expected := `[[a]]
|
||||||
[[a]]
|
|
||||||
[a.b]
|
[a.b]
|
||||||
c = 'd'
|
c = 'd'
|
||||||
`
|
`
|
||||||
equalStringsIgnoreNewlines(t, expected, buf.String())
|
assert.Equal(t, expected, buf.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestIssue424(t *testing.T) {
|
func TestIssue424(t *testing.T) {
|
||||||
@@ -782,6 +942,254 @@ func TestIssue424(t *testing.T) {
|
|||||||
require.Equal(t, msg2, msg2parsed)
|
require.Equal(t, msg2, msg2parsed)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestIssue567(t *testing.T) {
|
||||||
|
var m map[string]interface{}
|
||||||
|
err := toml.Unmarshal([]byte("A = 12:08:05"), &m)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.IsType(t, m["A"], toml.LocalTime{})
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestIssue590(t *testing.T) {
|
||||||
|
type CustomType int
|
||||||
|
var cfg struct {
|
||||||
|
Option CustomType `toml:"option"`
|
||||||
|
}
|
||||||
|
err := toml.Unmarshal([]byte("option = 42"), &cfg)
|
||||||
|
require.NoError(t, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestIssue571(t *testing.T) {
|
||||||
|
type Foo struct {
|
||||||
|
Float32 float32
|
||||||
|
Float64 float64
|
||||||
|
}
|
||||||
|
|
||||||
|
const closeEnough = 1e-9
|
||||||
|
|
||||||
|
foo := Foo{
|
||||||
|
Float32: 42,
|
||||||
|
Float64: 43,
|
||||||
|
}
|
||||||
|
b, err := toml.Marshal(foo)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
var foo2 Foo
|
||||||
|
err = toml.Unmarshal(b, &foo2)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
assert.InDelta(t, 42, foo2.Float32, closeEnough)
|
||||||
|
assert.InDelta(t, 43, foo2.Float64, closeEnough)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestIssue678(t *testing.T) {
|
||||||
|
type Config struct {
|
||||||
|
BigInt big.Int
|
||||||
|
}
|
||||||
|
|
||||||
|
cfg := &Config{
|
||||||
|
BigInt: *big.NewInt(123),
|
||||||
|
}
|
||||||
|
|
||||||
|
out, err := toml.Marshal(cfg)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, "BigInt = '123'\n", string(out))
|
||||||
|
|
||||||
|
cfg2 := &Config{}
|
||||||
|
err = toml.Unmarshal(out, cfg2)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, cfg, cfg2)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestIssue752(t *testing.T) {
|
||||||
|
type Fooer interface {
|
||||||
|
Foo() string
|
||||||
|
}
|
||||||
|
|
||||||
|
type Container struct {
|
||||||
|
Fooer
|
||||||
|
}
|
||||||
|
|
||||||
|
c := Container{}
|
||||||
|
|
||||||
|
out, err := toml.Marshal(c)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, "", string(out))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestIssue768(t *testing.T) {
|
||||||
|
type cfg struct {
|
||||||
|
Name string `comment:"This is a multiline comment.\nThis is line 2."`
|
||||||
|
}
|
||||||
|
|
||||||
|
out, err := toml.Marshal(&cfg{})
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
expected := `# This is a multiline comment.
|
||||||
|
# This is line 2.
|
||||||
|
Name = ''
|
||||||
|
`
|
||||||
|
|
||||||
|
require.Equal(t, expected, string(out))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestIssue786(t *testing.T) {
|
||||||
|
type Dependencies struct {
|
||||||
|
Dependencies []string `toml:"dependencies,multiline,omitempty"`
|
||||||
|
BuildDependencies []string `toml:"buildDependencies,multiline,omitempty"`
|
||||||
|
OptionalDependencies []string `toml:"optionalDependencies,multiline,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type Test struct {
|
||||||
|
Dependencies Dependencies `toml:"dependencies,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
x := Test{}
|
||||||
|
b, err := toml.Marshal(x)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
require.Equal(t, "", string(b))
|
||||||
|
|
||||||
|
type General struct {
|
||||||
|
From string `toml:"from,omitempty" json:"from,omitempty" comment:"from in graphite-web format, the local TZ is used"`
|
||||||
|
Randomize bool `toml:"randomize" json:"randomize" comment:"randomize starting time with [0,step)"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type Custom struct {
|
||||||
|
Name string `toml:"name" json:"name,omitempty" comment:"names for generator, braces are expanded like in shell"`
|
||||||
|
Type string `toml:"type,omitempty" json:"type,omitempty" comment:"type of generator"`
|
||||||
|
General
|
||||||
|
}
|
||||||
|
type Config struct {
|
||||||
|
General
|
||||||
|
Custom []Custom `toml:"custom,omitempty" json:"custom,omitempty" comment:"generators with custom parameters can be specified separately"`
|
||||||
|
}
|
||||||
|
|
||||||
|
buf := new(bytes.Buffer)
|
||||||
|
config := &Config{General: General{From: "-2d", Randomize: true}}
|
||||||
|
config.Custom = []Custom{{Name: "omit", General: General{Randomize: false}}}
|
||||||
|
config.Custom = append(config.Custom, Custom{Name: "present", General: General{From: "-2d", Randomize: true}})
|
||||||
|
encoder := toml.NewEncoder(buf)
|
||||||
|
encoder.Encode(config)
|
||||||
|
|
||||||
|
expected := `# from in graphite-web format, the local TZ is used
|
||||||
|
from = '-2d'
|
||||||
|
# randomize starting time with [0,step)
|
||||||
|
randomize = true
|
||||||
|
|
||||||
|
# generators with custom parameters can be specified separately
|
||||||
|
[[custom]]
|
||||||
|
# names for generator, braces are expanded like in shell
|
||||||
|
name = 'omit'
|
||||||
|
# randomize starting time with [0,step)
|
||||||
|
randomize = false
|
||||||
|
|
||||||
|
[[custom]]
|
||||||
|
# names for generator, braces are expanded like in shell
|
||||||
|
name = 'present'
|
||||||
|
# from in graphite-web format, the local TZ is used
|
||||||
|
from = '-2d'
|
||||||
|
# randomize starting time with [0,step)
|
||||||
|
randomize = true
|
||||||
|
`
|
||||||
|
|
||||||
|
require.Equal(t, expected, buf.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestMarshalNestedAnonymousStructs(t *testing.T) {
|
||||||
|
type Embedded struct {
|
||||||
|
Value string `toml:"value" json:"value"`
|
||||||
|
Top struct {
|
||||||
|
Value string `toml:"value" json:"value"`
|
||||||
|
} `toml:"top" json:"top"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type Named struct {
|
||||||
|
Value string `toml:"value" json:"value"`
|
||||||
|
}
|
||||||
|
|
||||||
|
var doc struct {
|
||||||
|
Embedded
|
||||||
|
Named `toml:"named" json:"named"`
|
||||||
|
Anonymous struct {
|
||||||
|
Value string `toml:"value" json:"value"`
|
||||||
|
} `toml:"anonymous" json:"anonymous"`
|
||||||
|
}
|
||||||
|
|
||||||
|
expected := `value = ''
|
||||||
|
|
||||||
|
[top]
|
||||||
|
value = ''
|
||||||
|
|
||||||
|
[named]
|
||||||
|
value = ''
|
||||||
|
|
||||||
|
[anonymous]
|
||||||
|
value = ''
|
||||||
|
`
|
||||||
|
|
||||||
|
result, err := toml.Marshal(doc)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, expected, string(result))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestMarshalNestedAnonymousStructs_DuplicateField(t *testing.T) {
|
||||||
|
type Embedded struct {
|
||||||
|
Value string `toml:"value" json:"value"`
|
||||||
|
Top struct {
|
||||||
|
Value string `toml:"value" json:"value"`
|
||||||
|
} `toml:"top" json:"top"`
|
||||||
|
}
|
||||||
|
|
||||||
|
var doc struct {
|
||||||
|
Value string `toml:"value" json:"value"`
|
||||||
|
Embedded
|
||||||
|
}
|
||||||
|
doc.Embedded.Value = "shadowed"
|
||||||
|
doc.Value = "shadows"
|
||||||
|
|
||||||
|
expected := `value = 'shadows'
|
||||||
|
|
||||||
|
[top]
|
||||||
|
value = ''
|
||||||
|
`
|
||||||
|
|
||||||
|
result, err := toml.Marshal(doc)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, expected, string(result))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestLocalTime(t *testing.T) {
|
||||||
|
v := map[string]toml.LocalTime{
|
||||||
|
"a": {
|
||||||
|
Hour: 1,
|
||||||
|
Minute: 2,
|
||||||
|
Second: 3,
|
||||||
|
Nanosecond: 4,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
expected := `a = 01:02:03.000000004
|
||||||
|
`
|
||||||
|
|
||||||
|
out, err := toml.Marshal(v)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, expected, string(out))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestMarshalUint64Overflow(t *testing.T) {
|
||||||
|
// The TOML spec only requires implementation to provide support for the
|
||||||
|
// int64 range. To avoid generating TOML documents that would not be
|
||||||
|
// supported by standard-compliant parsers, uint64 > max int64 cannot be
|
||||||
|
// marshaled.
|
||||||
|
x := map[string]interface{}{
|
||||||
|
"foo": uint64(math.MaxInt64) + 1,
|
||||||
|
}
|
||||||
|
|
||||||
|
_, err := toml.Marshal(x)
|
||||||
|
require.Error(t, err)
|
||||||
|
}
|
||||||
|
|
||||||
func ExampleMarshal() {
|
func ExampleMarshal() {
|
||||||
type MyConfig struct {
|
type MyConfig struct {
|
||||||
Version int
|
Version int
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ package toml
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"strconv"
|
"unicode"
|
||||||
|
|
||||||
"github.com/pelletier/go-toml/v2/internal/ast"
|
"github.com/pelletier/go-toml/v2/internal/ast"
|
||||||
"github.com/pelletier/go-toml/v2/internal/danger"
|
"github.com/pelletier/go-toml/v2/internal/danger"
|
||||||
@@ -28,6 +28,11 @@ func (p *parser) Raw(raw ast.Range) []byte {
|
|||||||
return p.data[raw.Offset : raw.Offset+raw.Length]
|
return p.data[raw.Offset : raw.Offset+raw.Length]
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (p *parser) SetRaw(ref ast.Reference, from []byte, to []byte) {
|
||||||
|
b := danger.BytesRange(from, to)
|
||||||
|
p.builder.NodeAt(ref).Raw = p.Range(b)
|
||||||
|
}
|
||||||
|
|
||||||
func (p *parser) Reset(b []byte) {
|
func (p *parser) Reset(b []byte) {
|
||||||
p.builder.Reset()
|
p.builder.Reset()
|
||||||
p.ref = ast.InvalidReference
|
p.ref = ast.InvalidReference
|
||||||
@@ -107,9 +112,8 @@ func (p *parser) parseExpression(b []byte) (ast.Reference, []byte, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if b[0] == '#' {
|
if b[0] == '#' {
|
||||||
_, rest := scanComment(b)
|
_, rest, err := scanComment(b)
|
||||||
|
return ref, rest, err
|
||||||
return ref, rest, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if b[0] == '\n' || b[0] == '\r' {
|
if b[0] == '\n' || b[0] == '\r' {
|
||||||
@@ -130,9 +134,8 @@ func (p *parser) parseExpression(b []byte) (ast.Reference, []byte, error) {
|
|||||||
b = p.parseWhitespace(b)
|
b = p.parseWhitespace(b)
|
||||||
|
|
||||||
if len(b) > 0 && b[0] == '#' {
|
if len(b) > 0 && b[0] == '#' {
|
||||||
_, rest := scanComment(b)
|
_, rest, err := scanComment(b)
|
||||||
|
return ref, rest, err
|
||||||
return ref, rest, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return ref, b, nil
|
return ref, b, nil
|
||||||
@@ -154,12 +157,14 @@ func (p *parser) parseArrayTable(b []byte) (ast.Reference, []byte, error) {
|
|||||||
ref := p.builder.Push(ast.Node{
|
ref := p.builder.Push(ast.Node{
|
||||||
Kind: ast.ArrayTable,
|
Kind: ast.ArrayTable,
|
||||||
})
|
})
|
||||||
|
start := b
|
||||||
|
|
||||||
b = b[2:]
|
b = b[2:]
|
||||||
b = p.parseWhitespace(b)
|
b = p.parseWhitespace(b)
|
||||||
|
|
||||||
k, b, err := p.parseKey(b)
|
k, b, err := p.parseKey(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
p.SetRaw(ref, start, b)
|
||||||
return ref, nil, err
|
return ref, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -168,11 +173,12 @@ func (p *parser) parseArrayTable(b []byte) (ast.Reference, []byte, error) {
|
|||||||
|
|
||||||
b, err = expect(']', b)
|
b, err = expect(']', b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
p.SetRaw(ref, start, b)
|
||||||
return ref, nil, err
|
return ref, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
b, err = expect(']', b)
|
b, err = expect(']', b)
|
||||||
|
p.SetRaw(ref, start, b)
|
||||||
return ref, b, err
|
return ref, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -183,12 +189,14 @@ func (p *parser) parseStdTable(b []byte) (ast.Reference, []byte, error) {
|
|||||||
ref := p.builder.Push(ast.Node{
|
ref := p.builder.Push(ast.Node{
|
||||||
Kind: ast.Table,
|
Kind: ast.Table,
|
||||||
})
|
})
|
||||||
|
start := b
|
||||||
|
|
||||||
b = b[1:]
|
b = b[1:]
|
||||||
b = p.parseWhitespace(b)
|
b = p.parseWhitespace(b)
|
||||||
|
|
||||||
key, b, err := p.parseKey(b)
|
key, b, err := p.parseKey(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
p.SetRaw(ref, start, b)
|
||||||
return ref, nil, err
|
return ref, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -198,6 +206,7 @@ func (p *parser) parseStdTable(b []byte) (ast.Reference, []byte, error) {
|
|||||||
|
|
||||||
b, err = expect(']', b)
|
b, err = expect(']', b)
|
||||||
|
|
||||||
|
p.SetRaw(ref, start, b)
|
||||||
return ref, b, err
|
return ref, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -206,10 +215,12 @@ func (p *parser) parseKeyval(b []byte) (ast.Reference, []byte, error) {
|
|||||||
ref := p.builder.Push(ast.Node{
|
ref := p.builder.Push(ast.Node{
|
||||||
Kind: ast.KeyValue,
|
Kind: ast.KeyValue,
|
||||||
})
|
})
|
||||||
|
start := b
|
||||||
|
|
||||||
key, b, err := p.parseKey(b)
|
key, b, err := p.parseKey(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return ast.InvalidReference, nil, err
|
p.SetRaw(ref, start, b)
|
||||||
|
return ast.InvalidReference, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// keyval-sep = ws %x3D ws ; =
|
// keyval-sep = ws %x3D ws ; =
|
||||||
@@ -217,24 +228,28 @@ func (p *parser) parseKeyval(b []byte) (ast.Reference, []byte, error) {
|
|||||||
b = p.parseWhitespace(b)
|
b = p.parseWhitespace(b)
|
||||||
|
|
||||||
if len(b) == 0 {
|
if len(b) == 0 {
|
||||||
return ast.InvalidReference, nil, newDecodeError(b, "expected = after a key, but the document ends there")
|
p.SetRaw(ref, start, b)
|
||||||
|
return ast.InvalidReference, b, newDecodeError(b, "expected = after a key, but the document ends there")
|
||||||
}
|
}
|
||||||
|
|
||||||
b, err = expect('=', b)
|
b, err = expect('=', b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return ast.InvalidReference, nil, err
|
p.SetRaw(ref, start, b)
|
||||||
|
return ast.InvalidReference, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
b = p.parseWhitespace(b)
|
b = p.parseWhitespace(b)
|
||||||
|
|
||||||
valRef, b, err := p.parseVal(b)
|
valRef, b, err := p.parseVal(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
p.SetRaw(ref, start, b)
|
||||||
return ref, b, err
|
return ref, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
p.builder.Chain(valRef, key)
|
p.builder.Chain(valRef, key)
|
||||||
p.builder.AttachChild(ref, valRef)
|
p.builder.AttachChild(ref, valRef)
|
||||||
|
|
||||||
|
p.SetRaw(ref, start, b)
|
||||||
return ref, b, err
|
return ref, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -244,7 +259,7 @@ func (p *parser) parseVal(b []byte) (ast.Reference, []byte, error) {
|
|||||||
ref := ast.InvalidReference
|
ref := ast.InvalidReference
|
||||||
|
|
||||||
if len(b) == 0 {
|
if len(b) == 0 {
|
||||||
return ref, nil, newDecodeError(b, "expected value, not eof")
|
return ref, b, newDecodeError(b, "expected value, not eof")
|
||||||
}
|
}
|
||||||
|
|
||||||
var err error
|
var err error
|
||||||
@@ -289,23 +304,25 @@ func (p *parser) parseVal(b []byte) (ast.Reference, []byte, error) {
|
|||||||
return ref, b, err
|
return ref, b, err
|
||||||
case 't':
|
case 't':
|
||||||
if !scanFollowsTrue(b) {
|
if !scanFollowsTrue(b) {
|
||||||
return ref, nil, newDecodeError(atmost(b, 4), "expected 'true'")
|
return ref, b, newDecodeError(atmost(b, 4), "expected 'true'")
|
||||||
}
|
}
|
||||||
|
|
||||||
ref = p.builder.Push(ast.Node{
|
ref = p.builder.Push(ast.Node{
|
||||||
Kind: ast.Bool,
|
Kind: ast.Bool,
|
||||||
Data: b[:4],
|
Data: b[:4],
|
||||||
|
Raw: p.Range(b[:4]),
|
||||||
})
|
})
|
||||||
|
|
||||||
return ref, b[4:], nil
|
return ref, b[4:], nil
|
||||||
case 'f':
|
case 'f':
|
||||||
if !scanFollowsFalse(b) {
|
if !scanFollowsFalse(b) {
|
||||||
return ref, nil, newDecodeError(atmost(b, 5), "expected 'false'")
|
return ref, b, newDecodeError(atmost(b, 5), "expected 'false'")
|
||||||
}
|
}
|
||||||
|
|
||||||
ref = p.builder.Push(ast.Node{
|
ref = p.builder.Push(ast.Node{
|
||||||
Kind: ast.Bool,
|
Kind: ast.Bool,
|
||||||
Data: b[:5],
|
Data: b[:5],
|
||||||
|
Raw: p.Range(b[:5]),
|
||||||
})
|
})
|
||||||
|
|
||||||
return ref, b[5:], nil
|
return ref, b[5:], nil
|
||||||
@@ -329,7 +346,7 @@ func atmost(b []byte, n int) []byte {
|
|||||||
func (p *parser) parseLiteralString(b []byte) ([]byte, []byte, []byte, error) {
|
func (p *parser) parseLiteralString(b []byte) ([]byte, []byte, []byte, error) {
|
||||||
v, rest, err := scanLiteralString(b)
|
v, rest, err := scanLiteralString(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, nil, err
|
return nil, nil, rest, err
|
||||||
}
|
}
|
||||||
|
|
||||||
return v, v[1 : len(v)-1], rest, nil
|
return v, v[1 : len(v)-1], rest, nil
|
||||||
@@ -344,6 +361,7 @@ func (p *parser) parseInlineTable(b []byte) (ast.Reference, []byte, error) {
|
|||||||
parent := p.builder.Push(ast.Node{
|
parent := p.builder.Push(ast.Node{
|
||||||
Kind: ast.InlineTable,
|
Kind: ast.InlineTable,
|
||||||
})
|
})
|
||||||
|
start := b
|
||||||
|
|
||||||
first := true
|
first := true
|
||||||
|
|
||||||
@@ -354,7 +372,14 @@ func (p *parser) parseInlineTable(b []byte) (ast.Reference, []byte, error) {
|
|||||||
var err error
|
var err error
|
||||||
|
|
||||||
for len(b) > 0 {
|
for len(b) > 0 {
|
||||||
|
previousB := b
|
||||||
b = p.parseWhitespace(b)
|
b = p.parseWhitespace(b)
|
||||||
|
|
||||||
|
if len(b) == 0 {
|
||||||
|
p.SetRaw(parent, start, b)
|
||||||
|
return parent, b, newDecodeError(previousB[:1], "inline table is incomplete")
|
||||||
|
}
|
||||||
|
|
||||||
if b[0] == '}' {
|
if b[0] == '}' {
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
@@ -362,7 +387,8 @@ func (p *parser) parseInlineTable(b []byte) (ast.Reference, []byte, error) {
|
|||||||
if !first {
|
if !first {
|
||||||
b, err = expect(',', b)
|
b, err = expect(',', b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return parent, nil, err
|
p.SetRaw(parent, start, b)
|
||||||
|
return parent, b, err
|
||||||
}
|
}
|
||||||
b = p.parseWhitespace(b)
|
b = p.parseWhitespace(b)
|
||||||
}
|
}
|
||||||
@@ -371,7 +397,8 @@ func (p *parser) parseInlineTable(b []byte) (ast.Reference, []byte, error) {
|
|||||||
|
|
||||||
kv, b, err = p.parseKeyval(b)
|
kv, b, err = p.parseKeyval(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return parent, nil, err
|
p.SetRaw(parent, start, b)
|
||||||
|
return parent, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
if first {
|
if first {
|
||||||
@@ -386,6 +413,7 @@ func (p *parser) parseInlineTable(b []byte) (ast.Reference, []byte, error) {
|
|||||||
|
|
||||||
rest, err := expect('}', b)
|
rest, err := expect('}', b)
|
||||||
|
|
||||||
|
p.SetRaw(parent, start, b)
|
||||||
return parent, rest, err
|
return parent, rest, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -398,6 +426,8 @@ func (p *parser) parseValArray(b []byte) (ast.Reference, []byte, error) {
|
|||||||
// array-values =/ ws-comment-newline val ws-comment-newline [ array-sep ]
|
// array-values =/ ws-comment-newline val ws-comment-newline [ array-sep ]
|
||||||
// array-sep = %x2C ; , Comma
|
// array-sep = %x2C ; , Comma
|
||||||
// ws-comment-newline = *( wschar / [ comment ] newline )
|
// ws-comment-newline = *( wschar / [ comment ] newline )
|
||||||
|
arrayStart := b
|
||||||
|
start := b
|
||||||
b = b[1:]
|
b = b[1:]
|
||||||
|
|
||||||
parent := p.builder.Push(ast.Node{
|
parent := p.builder.Push(ast.Node{
|
||||||
@@ -412,11 +442,13 @@ func (p *parser) parseValArray(b []byte) (ast.Reference, []byte, error) {
|
|||||||
for len(b) > 0 {
|
for len(b) > 0 {
|
||||||
b, err = p.parseOptionalWhitespaceCommentNewline(b)
|
b, err = p.parseOptionalWhitespaceCommentNewline(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return parent, nil, err
|
p.SetRaw(parent, start, b)
|
||||||
|
return parent, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(b) == 0 {
|
if len(b) == 0 {
|
||||||
return parent, nil, newDecodeError(b, "array is incomplete")
|
p.SetRaw(parent, start, b)
|
||||||
|
return parent, b, newDecodeError(arrayStart[:1], "array is incomplete")
|
||||||
}
|
}
|
||||||
|
|
||||||
if b[0] == ']' {
|
if b[0] == ']' {
|
||||||
@@ -425,14 +457,19 @@ func (p *parser) parseValArray(b []byte) (ast.Reference, []byte, error) {
|
|||||||
|
|
||||||
if b[0] == ',' {
|
if b[0] == ',' {
|
||||||
if first {
|
if first {
|
||||||
return parent, nil, newDecodeError(b[0:1], "array cannot start with comma")
|
p.SetRaw(parent, start, b)
|
||||||
|
return parent, b, newDecodeError(b[0:1], "array cannot start with comma")
|
||||||
}
|
}
|
||||||
b = b[1:]
|
b = b[1:]
|
||||||
|
|
||||||
b, err = p.parseOptionalWhitespaceCommentNewline(b)
|
b, err = p.parseOptionalWhitespaceCommentNewline(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return parent, nil, err
|
p.SetRaw(parent, start, b)
|
||||||
|
return parent, b, err
|
||||||
}
|
}
|
||||||
|
} else if !first {
|
||||||
|
p.SetRaw(parent, start, b)
|
||||||
|
return parent, b, newDecodeError(b[0:1], "array elements must be separated by commas")
|
||||||
}
|
}
|
||||||
|
|
||||||
// TOML allows trailing commas in arrays.
|
// TOML allows trailing commas in arrays.
|
||||||
@@ -441,10 +478,10 @@ func (p *parser) parseValArray(b []byte) (ast.Reference, []byte, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
var valueRef ast.Reference
|
var valueRef ast.Reference
|
||||||
|
|
||||||
valueRef, b, err = p.parseVal(b)
|
valueRef, b, err = p.parseVal(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return parent, nil, err
|
p.SetRaw(parent, start, b)
|
||||||
|
return parent, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
if first {
|
if first {
|
||||||
@@ -456,13 +493,16 @@ func (p *parser) parseValArray(b []byte) (ast.Reference, []byte, error) {
|
|||||||
|
|
||||||
b, err = p.parseOptionalWhitespaceCommentNewline(b)
|
b, err = p.parseOptionalWhitespaceCommentNewline(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return parent, nil, err
|
p.SetRaw(parent, start, b)
|
||||||
|
return parent, b, err
|
||||||
}
|
}
|
||||||
first = false
|
first = false
|
||||||
}
|
}
|
||||||
|
|
||||||
rest, err := expect(']', b)
|
rest, err := expect(']', b)
|
||||||
|
|
||||||
|
p.SetRaw(parent, start, rest)
|
||||||
|
|
||||||
return parent, rest, err
|
return parent, rest, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -472,7 +512,10 @@ func (p *parser) parseOptionalWhitespaceCommentNewline(b []byte) ([]byte, error)
|
|||||||
b = p.parseWhitespace(b)
|
b = p.parseWhitespace(b)
|
||||||
|
|
||||||
if len(b) > 0 && b[0] == '#' {
|
if len(b) > 0 && b[0] == '#' {
|
||||||
_, b = scanComment(b)
|
_, b, err = scanComment(b)
|
||||||
|
if err != nil {
|
||||||
|
return b, err
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(b) == 0 {
|
if len(b) == 0 {
|
||||||
@@ -482,7 +525,7 @@ func (p *parser) parseOptionalWhitespaceCommentNewline(b []byte) ([]byte, error)
|
|||||||
if b[0] == '\n' || b[0] == '\r' {
|
if b[0] == '\n' || b[0] == '\r' {
|
||||||
b, err = p.parseNewline(b)
|
b, err = p.parseNewline(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return b, err
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
break
|
break
|
||||||
@@ -495,7 +538,7 @@ func (p *parser) parseOptionalWhitespaceCommentNewline(b []byte) ([]byte, error)
|
|||||||
func (p *parser) parseMultilineLiteralString(b []byte) ([]byte, []byte, []byte, error) {
|
func (p *parser) parseMultilineLiteralString(b []byte) ([]byte, []byte, []byte, error) {
|
||||||
token, rest, err := scanMultilineLiteralString(b)
|
token, rest, err := scanMultilineLiteralString(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, nil, err
|
return nil, nil, rest, err
|
||||||
}
|
}
|
||||||
|
|
||||||
i := 3
|
i := 3
|
||||||
@@ -522,9 +565,9 @@ func (p *parser) parseMultilineBasicString(b []byte) ([]byte, []byte, []byte, er
|
|||||||
// mlb-quotes = 1*2quotation-mark
|
// mlb-quotes = 1*2quotation-mark
|
||||||
// mlb-unescaped = wschar / %x21 / %x23-5B / %x5D-7E / non-ascii
|
// mlb-unescaped = wschar / %x21 / %x23-5B / %x5D-7E / non-ascii
|
||||||
// mlb-escaped-nl = escape ws newline *( wschar / newline )
|
// mlb-escaped-nl = escape ws newline *( wschar / newline )
|
||||||
token, rest, err := scanMultilineBasicString(b)
|
token, escaped, rest, err := scanMultilineBasicString(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, nil, err
|
return nil, nil, rest, err
|
||||||
}
|
}
|
||||||
|
|
||||||
i := 3
|
i := 3
|
||||||
@@ -539,21 +582,21 @@ func (p *parser) parseMultilineBasicString(b []byte) ([]byte, []byte, []byte, er
|
|||||||
// fast path
|
// fast path
|
||||||
startIdx := i
|
startIdx := i
|
||||||
endIdx := len(token) - len(`"""`)
|
endIdx := len(token) - len(`"""`)
|
||||||
for ; i < endIdx; i++ {
|
|
||||||
if token[i] == '\\' {
|
if !escaped {
|
||||||
break
|
str := token[startIdx:endIdx]
|
||||||
|
verr := utf8TomlValidAlreadyEscaped(str)
|
||||||
|
if verr.Zero() {
|
||||||
|
return token, str, rest, nil
|
||||||
}
|
}
|
||||||
}
|
return nil, nil, rest, newDecodeError(str[verr.Index:verr.Index+verr.Size], "invalid UTF-8")
|
||||||
if i == endIdx {
|
|
||||||
return token, token[startIdx:endIdx], rest, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
var builder bytes.Buffer
|
var builder bytes.Buffer
|
||||||
builder.Write(token[startIdx:i])
|
|
||||||
|
|
||||||
// The scanner ensures that the token starts and ends with quotes and that
|
// The scanner ensures that the token starts and ends with quotes and that
|
||||||
// escapes are balanced.
|
// escapes are balanced.
|
||||||
for ; i < len(token)-3; i++ {
|
for i < len(token)-3 {
|
||||||
c := token[i]
|
c := token[i]
|
||||||
|
|
||||||
//nolint:nestif
|
//nolint:nestif
|
||||||
@@ -561,17 +604,33 @@ func (p *parser) parseMultilineBasicString(b []byte) ([]byte, []byte, []byte, er
|
|||||||
// When the last non-whitespace character on a line is an unescaped \,
|
// When the last non-whitespace character on a line is an unescaped \,
|
||||||
// it will be trimmed along with all whitespace (including newlines) up
|
// it will be trimmed along with all whitespace (including newlines) up
|
||||||
// to the next non-whitespace character or closing delimiter.
|
// to the next non-whitespace character or closing delimiter.
|
||||||
if token[i+1] == '\n' || (token[i+1] == '\r' && token[i+2] == '\n') {
|
|
||||||
i++ // skip the \
|
isLastNonWhitespaceOnLine := false
|
||||||
|
j := 1
|
||||||
|
findEOLLoop:
|
||||||
|
for ; j < len(token)-3-i; j++ {
|
||||||
|
switch token[i+j] {
|
||||||
|
case ' ', '\t':
|
||||||
|
continue
|
||||||
|
case '\r':
|
||||||
|
if token[i+j+1] == '\n' {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
case '\n':
|
||||||
|
isLastNonWhitespaceOnLine = true
|
||||||
|
}
|
||||||
|
break findEOLLoop
|
||||||
|
}
|
||||||
|
if isLastNonWhitespaceOnLine {
|
||||||
|
i += j
|
||||||
for ; i < len(token)-3; i++ {
|
for ; i < len(token)-3; i++ {
|
||||||
c := token[i]
|
c := token[i]
|
||||||
if !(c == '\n' || c == '\r' || c == ' ' || c == '\t') {
|
if !(c == '\n' || c == '\r' || c == ' ' || c == '\t') {
|
||||||
i--
|
i--
|
||||||
|
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
i++
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -592,27 +651,34 @@ func (p *parser) parseMultilineBasicString(b []byte) ([]byte, []byte, []byte, er
|
|||||||
builder.WriteByte('\r')
|
builder.WriteByte('\r')
|
||||||
case 't':
|
case 't':
|
||||||
builder.WriteByte('\t')
|
builder.WriteByte('\t')
|
||||||
|
case 'e':
|
||||||
|
builder.WriteByte(0x1B)
|
||||||
case 'u':
|
case 'u':
|
||||||
x, err := hexToString(atmost(token[i+1:], 4), 4)
|
x, err := hexToRune(atmost(token[i+1:], 4), 4)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, nil, err
|
return nil, nil, rest, err
|
||||||
}
|
}
|
||||||
|
builder.WriteRune(x)
|
||||||
builder.WriteString(x)
|
|
||||||
i += 4
|
i += 4
|
||||||
case 'U':
|
case 'U':
|
||||||
x, err := hexToString(atmost(token[i+1:], 8), 8)
|
x, err := hexToRune(atmost(token[i+1:], 8), 8)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, nil, err
|
return nil, nil, rest, err
|
||||||
}
|
}
|
||||||
|
|
||||||
builder.WriteString(x)
|
builder.WriteRune(x)
|
||||||
i += 8
|
i += 8
|
||||||
default:
|
default:
|
||||||
return nil, nil, nil, newDecodeError(token[i:i+1], "invalid escaped character %#U", c)
|
return nil, nil, rest, newDecodeError(token[i:i+1], "invalid escaped character %#U", c)
|
||||||
}
|
}
|
||||||
|
i++
|
||||||
} else {
|
} else {
|
||||||
builder.WriteByte(c)
|
size := utf8ValidNext(token[i:])
|
||||||
|
if size == 0 {
|
||||||
|
return nil, nil, rest, newDecodeError(token[i:i+1], "invalid character %#U", c)
|
||||||
|
}
|
||||||
|
builder.Write(token[i : i+size])
|
||||||
|
i += size
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -630,7 +696,7 @@ func (p *parser) parseKey(b []byte) (ast.Reference, []byte, error) {
|
|||||||
// dot-sep = ws %x2E ws ; . Period
|
// dot-sep = ws %x2E ws ; . Period
|
||||||
raw, key, b, err := p.parseSimpleKey(b)
|
raw, key, b, err := p.parseSimpleKey(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return ast.InvalidReference, nil, err
|
return ast.InvalidReference, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
ref := p.builder.Push(ast.Node{
|
ref := p.builder.Push(ast.Node{
|
||||||
@@ -646,7 +712,7 @@ func (p *parser) parseKey(b []byte) (ast.Reference, []byte, error) {
|
|||||||
|
|
||||||
raw, key, b, err = p.parseSimpleKey(b)
|
raw, key, b, err = p.parseSimpleKey(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return ref, nil, err
|
return ref, b, err
|
||||||
}
|
}
|
||||||
|
|
||||||
p.builder.PushAndChain(ast.Node{
|
p.builder.PushAndChain(ast.Node{
|
||||||
@@ -663,13 +729,13 @@ func (p *parser) parseKey(b []byte) (ast.Reference, []byte, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (p *parser) parseSimpleKey(b []byte) (raw, key, rest []byte, err error) {
|
func (p *parser) parseSimpleKey(b []byte) (raw, key, rest []byte, err error) {
|
||||||
|
if len(b) == 0 {
|
||||||
|
return nil, nil, b, newDecodeError(b, "expected key but found none")
|
||||||
|
}
|
||||||
|
|
||||||
// simple-key = quoted-key / unquoted-key
|
// simple-key = quoted-key / unquoted-key
|
||||||
// unquoted-key = 1*( ALPHA / DIGIT / %x2D / %x5F ) ; A-Z / a-z / 0-9 / - / _
|
// unquoted-key = 1*( ALPHA / DIGIT / %x2D / %x5F ) ; A-Z / a-z / 0-9 / - / _
|
||||||
// quoted-key = basic-string / literal-string
|
// quoted-key = basic-string / literal-string
|
||||||
if len(b) == 0 {
|
|
||||||
return nil, nil, nil, newDecodeError(b, "key is incomplete")
|
|
||||||
}
|
|
||||||
|
|
||||||
switch {
|
switch {
|
||||||
case b[0] == '\'':
|
case b[0] == '\'':
|
||||||
return p.parseLiteralString(b)
|
return p.parseLiteralString(b)
|
||||||
@@ -679,7 +745,7 @@ func (p *parser) parseSimpleKey(b []byte) (raw, key, rest []byte, err error) {
|
|||||||
key, rest = scanUnquotedKey(b)
|
key, rest = scanUnquotedKey(b)
|
||||||
return key, key, rest, nil
|
return key, key, rest, nil
|
||||||
default:
|
default:
|
||||||
return nil, nil, nil, newDecodeError(b[0:1], "invalid character at start of key: %c", b[0])
|
return nil, nil, b[1:], newDecodeError(b[0:1], "invalid character at start of key: %c", b[0])
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -699,30 +765,33 @@ func (p *parser) parseBasicString(b []byte) ([]byte, []byte, []byte, error) {
|
|||||||
// escape-seq-char =/ %x74 ; t tab U+0009
|
// escape-seq-char =/ %x74 ; t tab U+0009
|
||||||
// escape-seq-char =/ %x75 4HEXDIG ; uXXXX U+XXXX
|
// escape-seq-char =/ %x75 4HEXDIG ; uXXXX U+XXXX
|
||||||
// escape-seq-char =/ %x55 8HEXDIG ; UXXXXXXXX U+XXXXXXXX
|
// escape-seq-char =/ %x55 8HEXDIG ; UXXXXXXXX U+XXXXXXXX
|
||||||
token, rest, err := scanBasicString(b)
|
token, escaped, rest, err := scanBasicString(b)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, nil, err
|
return nil, nil, rest, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// fast path
|
startIdx := len(`"`)
|
||||||
i := len(`"`)
|
|
||||||
startIdx := i
|
|
||||||
endIdx := len(token) - len(`"`)
|
endIdx := len(token) - len(`"`)
|
||||||
for ; i < endIdx; i++ {
|
|
||||||
if token[i] == '\\' {
|
// Fast path. If there is no escape sequence, the string should just be
|
||||||
break
|
// an UTF-8 encoded string, which is the same as Go. In that case,
|
||||||
|
// validate the string and return a direct reference to the buffer.
|
||||||
|
if !escaped {
|
||||||
|
str := token[startIdx:endIdx]
|
||||||
|
verr := utf8TomlValidAlreadyEscaped(str)
|
||||||
|
if verr.Zero() {
|
||||||
|
return token, str, rest, nil
|
||||||
}
|
}
|
||||||
|
return nil, nil, rest, newDecodeError(str[verr.Index:verr.Index+verr.Size], "invalid UTF-8")
|
||||||
}
|
}
|
||||||
if i == endIdx {
|
|
||||||
return token, token[startIdx:endIdx], rest, nil
|
i := startIdx
|
||||||
}
|
|
||||||
|
|
||||||
var builder bytes.Buffer
|
var builder bytes.Buffer
|
||||||
builder.Write(token[startIdx:i])
|
|
||||||
|
|
||||||
// The scanner ensures that the token starts and ends with quotes and that
|
// The scanner ensures that the token starts and ends with quotes and that
|
||||||
// escapes are balanced.
|
// escapes are balanced.
|
||||||
for ; i < len(token)-1; i++ {
|
for i < len(token)-1 {
|
||||||
c := token[i]
|
c := token[i]
|
||||||
if c == '\\' {
|
if c == '\\' {
|
||||||
i++
|
i++
|
||||||
@@ -741,47 +810,68 @@ func (p *parser) parseBasicString(b []byte) ([]byte, []byte, []byte, error) {
|
|||||||
builder.WriteByte('\r')
|
builder.WriteByte('\r')
|
||||||
case 't':
|
case 't':
|
||||||
builder.WriteByte('\t')
|
builder.WriteByte('\t')
|
||||||
|
case 'e':
|
||||||
|
builder.WriteByte(0x1B)
|
||||||
case 'u':
|
case 'u':
|
||||||
x, err := hexToString(token[i+1:len(token)-1], 4)
|
x, err := hexToRune(token[i+1:len(token)-1], 4)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, nil, err
|
return nil, nil, rest, err
|
||||||
}
|
}
|
||||||
|
|
||||||
builder.WriteString(x)
|
builder.WriteRune(x)
|
||||||
i += 4
|
i += 4
|
||||||
case 'U':
|
case 'U':
|
||||||
x, err := hexToString(token[i+1:len(token)-1], 8)
|
x, err := hexToRune(token[i+1:len(token)-1], 8)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, nil, err
|
return nil, nil, rest, err
|
||||||
}
|
}
|
||||||
|
|
||||||
builder.WriteString(x)
|
builder.WriteRune(x)
|
||||||
i += 8
|
i += 8
|
||||||
default:
|
default:
|
||||||
return nil, nil, nil, newDecodeError(token[i:i+1], "invalid escaped character %#U", c)
|
return nil, nil, rest, newDecodeError(token[i:i+1], "invalid escaped character %#U", c)
|
||||||
}
|
}
|
||||||
|
i++
|
||||||
} else {
|
} else {
|
||||||
builder.WriteByte(c)
|
size := utf8ValidNext(token[i:])
|
||||||
|
if size == 0 {
|
||||||
|
return nil, nil, rest, newDecodeError(token[i:i+1], "invalid character %#U", c)
|
||||||
|
}
|
||||||
|
builder.Write(token[i : i+size])
|
||||||
|
i += size
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return token, builder.Bytes(), rest, nil
|
return token, builder.Bytes(), rest, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func hexToString(b []byte, length int) (string, error) {
|
func hexToRune(b []byte, length int) (rune, error) {
|
||||||
if len(b) < length {
|
if len(b) < length {
|
||||||
return "", newDecodeError(b, "unicode point needs %d character, not %d", length, len(b))
|
return -1, newDecodeError(b, "unicode point needs %d character, not %d", length, len(b))
|
||||||
}
|
}
|
||||||
b = b[:length]
|
b = b[:length]
|
||||||
|
|
||||||
//nolint:godox
|
var r uint32
|
||||||
// TODO: slow
|
for i, c := range b {
|
||||||
intcode, err := strconv.ParseInt(string(b), 16, 32)
|
d := uint32(0)
|
||||||
if err != nil {
|
switch {
|
||||||
return "", newDecodeError(b, "couldn't parse hexadecimal number: %w", err)
|
case '0' <= c && c <= '9':
|
||||||
|
d = uint32(c - '0')
|
||||||
|
case 'a' <= c && c <= 'f':
|
||||||
|
d = uint32(c - 'a' + 10)
|
||||||
|
case 'A' <= c && c <= 'F':
|
||||||
|
d = uint32(c - 'A' + 10)
|
||||||
|
default:
|
||||||
|
return -1, newDecodeError(b[i:i+1], "non-hex character")
|
||||||
|
}
|
||||||
|
r = r*16 + d
|
||||||
}
|
}
|
||||||
|
|
||||||
return string(rune(intcode)), nil
|
if r > unicode.MaxRune || 0xD800 <= r && r < 0xE000 {
|
||||||
|
return -1, newDecodeError(b, "escape sequence is invalid Unicode code point")
|
||||||
|
}
|
||||||
|
|
||||||
|
return rune(r), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *parser) parseWhitespace(b []byte) []byte {
|
func (p *parser) parseWhitespace(b []byte) []byte {
|
||||||
@@ -798,27 +888,28 @@ func (p *parser) parseIntOrFloatOrDateTime(b []byte) (ast.Reference, []byte, err
|
|||||||
switch b[0] {
|
switch b[0] {
|
||||||
case 'i':
|
case 'i':
|
||||||
if !scanFollowsInf(b) {
|
if !scanFollowsInf(b) {
|
||||||
return ast.InvalidReference, nil, newDecodeError(atmost(b, 3), "expected 'inf'")
|
return ast.InvalidReference, b, newDecodeError(atmost(b, 3), "expected 'inf'")
|
||||||
}
|
}
|
||||||
|
|
||||||
return p.builder.Push(ast.Node{
|
return p.builder.Push(ast.Node{
|
||||||
Kind: ast.Float,
|
Kind: ast.Float,
|
||||||
Data: b[:3],
|
Data: b[:3],
|
||||||
|
Raw: p.Range(b[:3]),
|
||||||
}), b[3:], nil
|
}), b[3:], nil
|
||||||
case 'n':
|
case 'n':
|
||||||
if !scanFollowsNan(b) {
|
if !scanFollowsNan(b) {
|
||||||
return ast.InvalidReference, nil, newDecodeError(atmost(b, 3), "expected 'nan'")
|
return ast.InvalidReference, b, newDecodeError(atmost(b, 3), "expected 'nan'")
|
||||||
}
|
}
|
||||||
|
|
||||||
return p.builder.Push(ast.Node{
|
return p.builder.Push(ast.Node{
|
||||||
Kind: ast.Float,
|
Kind: ast.Float,
|
||||||
Data: b[:3],
|
Data: b[:3],
|
||||||
|
Raw: p.Range(b[:3]),
|
||||||
}), b[3:], nil
|
}), b[3:], nil
|
||||||
case '+', '-':
|
case '+', '-':
|
||||||
return p.scanIntOrFloat(b)
|
return p.scanIntOrFloat(b)
|
||||||
}
|
}
|
||||||
|
|
||||||
//nolint:gomnd
|
|
||||||
if len(b) < 3 {
|
if len(b) < 3 {
|
||||||
return p.scanIntOrFloat(b)
|
return p.scanIntOrFloat(b)
|
||||||
}
|
}
|
||||||
@@ -836,26 +927,17 @@ func (p *parser) parseIntOrFloatOrDateTime(b []byte) (ast.Reference, []byte, err
|
|||||||
if idx == 2 && c == ':' || (idx == 4 && c == '-') {
|
if idx == 2 && c == ':' || (idx == 4 && c == '-') {
|
||||||
return p.scanDateTime(b)
|
return p.scanDateTime(b)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
return p.scanIntOrFloat(b)
|
return p.scanIntOrFloat(b)
|
||||||
}
|
}
|
||||||
|
|
||||||
func digitsToInt(b []byte) int {
|
|
||||||
x := 0
|
|
||||||
|
|
||||||
for _, d := range b {
|
|
||||||
x *= 10
|
|
||||||
x += int(d - '0')
|
|
||||||
}
|
|
||||||
|
|
||||||
return x
|
|
||||||
}
|
|
||||||
|
|
||||||
//nolint:gocognit,cyclop
|
|
||||||
func (p *parser) scanDateTime(b []byte) (ast.Reference, []byte, error) {
|
func (p *parser) scanDateTime(b []byte) (ast.Reference, []byte, error) {
|
||||||
// scans for contiguous characters in [0-9T:Z.+-], and up to one space if
|
// scans for contiguous characters in [0-9T:Z.+-], and up to one space if
|
||||||
// followed by a digit.
|
// followed by a digit.
|
||||||
|
hasDate := false
|
||||||
hasTime := false
|
hasTime := false
|
||||||
hasTz := false
|
hasTz := false
|
||||||
seenSpace := false
|
seenSpace := false
|
||||||
@@ -868,17 +950,23 @@ byteLoop:
|
|||||||
switch {
|
switch {
|
||||||
case isDigit(c):
|
case isDigit(c):
|
||||||
case c == '-':
|
case c == '-':
|
||||||
|
hasDate = true
|
||||||
const minOffsetOfTz = 8
|
const minOffsetOfTz = 8
|
||||||
if i >= minOffsetOfTz {
|
if i >= minOffsetOfTz {
|
||||||
hasTz = true
|
hasTz = true
|
||||||
}
|
}
|
||||||
case c == 'T' || c == ':' || c == '.':
|
case c == 'T' || c == 't' || c == ':' || c == '.':
|
||||||
hasTime = true
|
hasTime = true
|
||||||
case c == '+' || c == '-' || c == 'Z':
|
case c == '+' || c == '-' || c == 'Z' || c == 'z':
|
||||||
hasTz = true
|
hasTz = true
|
||||||
case c == ' ':
|
case c == ' ':
|
||||||
if !seenSpace && i+1 < len(b) && isDigit(b[i+1]) {
|
if !seenSpace && i+1 < len(b) && isDigit(b[i+1]) {
|
||||||
i += 2
|
i += 2
|
||||||
|
// Avoid reaching past the end of the document in case the time
|
||||||
|
// is malformed. See TestIssue585.
|
||||||
|
if i >= len(b) {
|
||||||
|
i--
|
||||||
|
}
|
||||||
seenSpace = true
|
seenSpace = true
|
||||||
hasTime = true
|
hasTime = true
|
||||||
} else {
|
} else {
|
||||||
@@ -892,10 +980,14 @@ byteLoop:
|
|||||||
var kind ast.Kind
|
var kind ast.Kind
|
||||||
|
|
||||||
if hasTime {
|
if hasTime {
|
||||||
if hasTz {
|
if hasDate {
|
||||||
kind = ast.DateTime
|
if hasTz {
|
||||||
|
kind = ast.DateTime
|
||||||
|
} else {
|
||||||
|
kind = ast.LocalDateTime
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
kind = ast.LocalDateTime
|
kind = ast.LocalTime
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
kind = ast.LocalDate
|
kind = ast.LocalDate
|
||||||
@@ -904,6 +996,7 @@ byteLoop:
|
|||||||
return p.builder.Push(ast.Node{
|
return p.builder.Push(ast.Node{
|
||||||
Kind: kind,
|
Kind: kind,
|
||||||
Data: b[:i],
|
Data: b[:i],
|
||||||
|
Raw: p.Range(b[:i]),
|
||||||
}), b[i:], nil
|
}), b[i:], nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -911,7 +1004,7 @@ byteLoop:
|
|||||||
func (p *parser) scanIntOrFloat(b []byte) (ast.Reference, []byte, error) {
|
func (p *parser) scanIntOrFloat(b []byte) (ast.Reference, []byte, error) {
|
||||||
i := 0
|
i := 0
|
||||||
|
|
||||||
if len(b) > 2 && b[0] == '0' && b[1] != '.' {
|
if len(b) > 2 && b[0] == '0' && b[1] != '.' && b[1] != 'e' && b[1] != 'E' {
|
||||||
var isValidRune validRuneFn
|
var isValidRune validRuneFn
|
||||||
|
|
||||||
switch b[1] {
|
switch b[1] {
|
||||||
@@ -937,6 +1030,7 @@ func (p *parser) scanIntOrFloat(b []byte) (ast.Reference, []byte, error) {
|
|||||||
return p.builder.Push(ast.Node{
|
return p.builder.Push(ast.Node{
|
||||||
Kind: ast.Integer,
|
Kind: ast.Integer,
|
||||||
Data: b[:i],
|
Data: b[:i],
|
||||||
|
Raw: p.Range(b[:i]),
|
||||||
}), b[i:], nil
|
}), b[i:], nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -960,10 +1054,11 @@ func (p *parser) scanIntOrFloat(b []byte) (ast.Reference, []byte, error) {
|
|||||||
return p.builder.Push(ast.Node{
|
return p.builder.Push(ast.Node{
|
||||||
Kind: ast.Float,
|
Kind: ast.Float,
|
||||||
Data: b[:i+3],
|
Data: b[:i+3],
|
||||||
|
Raw: p.Range(b[:i+3]),
|
||||||
}), b[i+3:], nil
|
}), b[i+3:], nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return ast.InvalidReference, nil, newDecodeError(b[i:i+1], "unexpected character 'i' while scanning for a number")
|
return ast.InvalidReference, b[i:], newDecodeError(b[i:i+1], "unexpected character 'i' while scanning for a number")
|
||||||
}
|
}
|
||||||
|
|
||||||
if c == 'n' {
|
if c == 'n' {
|
||||||
@@ -971,10 +1066,11 @@ func (p *parser) scanIntOrFloat(b []byte) (ast.Reference, []byte, error) {
|
|||||||
return p.builder.Push(ast.Node{
|
return p.builder.Push(ast.Node{
|
||||||
Kind: ast.Float,
|
Kind: ast.Float,
|
||||||
Data: b[:i+3],
|
Data: b[:i+3],
|
||||||
|
Raw: p.Range(b[:i+3]),
|
||||||
}), b[i+3:], nil
|
}), b[i+3:], nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return ast.InvalidReference, nil, newDecodeError(b[i:i+1], "unexpected character 'n' while scanning for a number")
|
return ast.InvalidReference, b[i:], newDecodeError(b[i:i+1], "unexpected character 'n' while scanning for a number")
|
||||||
}
|
}
|
||||||
|
|
||||||
break
|
break
|
||||||
@@ -993,6 +1089,7 @@ func (p *parser) scanIntOrFloat(b []byte) (ast.Reference, []byte, error) {
|
|||||||
return p.builder.Push(ast.Node{
|
return p.builder.Push(ast.Node{
|
||||||
Kind: kind,
|
Kind: kind,
|
||||||
Data: b[:i],
|
Data: b[:i],
|
||||||
|
Raw: p.Range(b[:i]),
|
||||||
}), b[i:], nil
|
}), b[i:], nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1018,8 +1115,12 @@ func isValidBinaryRune(r byte) bool {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func expect(x byte, b []byte) ([]byte, error) {
|
func expect(x byte, b []byte) ([]byte, error) {
|
||||||
|
if len(b) == 0 {
|
||||||
|
return b, newDecodeError(b, "expected character %c but the document ended here", x)
|
||||||
|
}
|
||||||
|
|
||||||
if b[0] != x {
|
if b[0] != x {
|
||||||
return nil, newDecodeError(b[0:1], "expected character %U", x)
|
return b, newDecodeError(b[0:1], "expected character %c", x)
|
||||||
}
|
}
|
||||||
|
|
||||||
return b[1:], nil
|
return b[1:], nil
|
||||||
|
|||||||
+100
@@ -1,6 +1,8 @@
|
|||||||
package toml
|
package toml
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/pelletier/go-toml/v2/internal/ast"
|
"github.com/pelletier/go-toml/v2/internal/ast"
|
||||||
@@ -348,3 +350,101 @@ func TestParser_AST(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func BenchmarkParseBasicStringWithUnicode(b *testing.B) {
|
||||||
|
p := &parser{}
|
||||||
|
b.Run("4", func(b *testing.B) {
|
||||||
|
input := []byte(`"\u1234\u5678\u9ABC\u1234\u5678\u9ABC"`)
|
||||||
|
b.ReportAllocs()
|
||||||
|
b.SetBytes(int64(len(input)))
|
||||||
|
|
||||||
|
for i := 0; i < b.N; i++ {
|
||||||
|
p.parseBasicString(input)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
b.Run("8", func(b *testing.B) {
|
||||||
|
input := []byte(`"\u12345678\u9ABCDEF0\u12345678\u9ABCDEF0"`)
|
||||||
|
b.ReportAllocs()
|
||||||
|
b.SetBytes(int64(len(input)))
|
||||||
|
|
||||||
|
for i := 0; i < b.N; i++ {
|
||||||
|
p.parseBasicString(input)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func BenchmarkParseBasicStringsEasy(b *testing.B) {
|
||||||
|
p := &parser{}
|
||||||
|
|
||||||
|
for _, size := range []int{1, 4, 8, 16, 21} {
|
||||||
|
b.Run(strconv.Itoa(size), func(b *testing.B) {
|
||||||
|
input := []byte(`"` + strings.Repeat("A", size) + `"`)
|
||||||
|
|
||||||
|
b.ReportAllocs()
|
||||||
|
b.SetBytes(int64(len(input)))
|
||||||
|
|
||||||
|
for i := 0; i < b.N; i++ {
|
||||||
|
p.parseBasicString(input)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestParser_AST_DateTimes(t *testing.T) {
|
||||||
|
examples := []struct {
|
||||||
|
desc string
|
||||||
|
input string
|
||||||
|
kind ast.Kind
|
||||||
|
err bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
desc: "offset-date-time with delim 'T' and UTC offset",
|
||||||
|
input: `2021-07-21T12:08:05Z`,
|
||||||
|
kind: ast.DateTime,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
desc: "offset-date-time with space delim and +8hours offset",
|
||||||
|
input: `2021-07-21 12:08:05+08:00`,
|
||||||
|
kind: ast.DateTime,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
desc: "local-date-time with nano second",
|
||||||
|
input: `2021-07-21T12:08:05.666666666`,
|
||||||
|
kind: ast.LocalDateTime,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
desc: "local-date-time",
|
||||||
|
input: `2021-07-21T12:08:05`,
|
||||||
|
kind: ast.LocalDateTime,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
desc: "local-date",
|
||||||
|
input: `2021-07-21`,
|
||||||
|
kind: ast.LocalDate,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, e := range examples {
|
||||||
|
e := e
|
||||||
|
t.Run(e.desc, func(t *testing.T) {
|
||||||
|
p := parser{}
|
||||||
|
p.Reset([]byte(`A = ` + e.input))
|
||||||
|
p.NextExpression()
|
||||||
|
err := p.Error()
|
||||||
|
if e.err {
|
||||||
|
require.Error(t, err)
|
||||||
|
} else {
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
expected := astNode{
|
||||||
|
Kind: ast.KeyValue,
|
||||||
|
Children: []astNode{
|
||||||
|
{Kind: e.kind, Data: []byte(e.input)},
|
||||||
|
{Kind: ast.Key, Data: []byte(`A`)},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
compareNode(t, expected, p.Expression())
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|||||||
+122
-26
@@ -49,16 +49,21 @@ func scanLiteralString(b []byte) ([]byte, []byte, error) {
|
|||||||
// literal-string = apostrophe *literal-char apostrophe
|
// literal-string = apostrophe *literal-char apostrophe
|
||||||
// apostrophe = %x27 ; ' apostrophe
|
// apostrophe = %x27 ; ' apostrophe
|
||||||
// literal-char = %x09 / %x20-26 / %x28-7E / non-ascii
|
// literal-char = %x09 / %x20-26 / %x28-7E / non-ascii
|
||||||
for i := 1; i < len(b); i++ {
|
for i := 1; i < len(b); {
|
||||||
switch b[i] {
|
switch b[i] {
|
||||||
case '\'':
|
case '\'':
|
||||||
return b[:i+1], b[i+1:], nil
|
return b[:i+1], b[i+1:], nil
|
||||||
case '\n':
|
case '\n', '\r':
|
||||||
return nil, nil, newDecodeError(b[i:i+1], "literal strings cannot have new lines")
|
return nil, b[i+1:], newDecodeError(b[i:i+1], "literal strings cannot have new lines")
|
||||||
}
|
}
|
||||||
|
size := utf8ValidNext(b[i:])
|
||||||
|
if size == 0 {
|
||||||
|
return nil, b[i+1:], newDecodeError(b[i:i+1], "invalid character")
|
||||||
|
}
|
||||||
|
i += size
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil, nil, newDecodeError(b[len(b):], "unterminated literal string")
|
return nil, b[len(b):], newDecodeError(b[len(b):], "unterminated literal string")
|
||||||
}
|
}
|
||||||
|
|
||||||
func scanMultilineLiteralString(b []byte) ([]byte, []byte, error) {
|
func scanMultilineLiteralString(b []byte) ([]byte, []byte, error) {
|
||||||
@@ -70,23 +75,62 @@ func scanMultilineLiteralString(b []byte) ([]byte, []byte, error) {
|
|||||||
// mll-content = mll-char / newline
|
// mll-content = mll-char / newline
|
||||||
// mll-char = %x09 / %x20-26 / %x28-7E / non-ascii
|
// mll-char = %x09 / %x20-26 / %x28-7E / non-ascii
|
||||||
// mll-quotes = 1*2apostrophe
|
// mll-quotes = 1*2apostrophe
|
||||||
for i := 3; i < len(b); i++ {
|
for i := 3; i < len(b); {
|
||||||
if b[i] == '\'' && scanFollowsMultilineLiteralStringDelimiter(b[i:]) {
|
switch b[i] {
|
||||||
return b[:i+3], b[i+3:], nil
|
case '\'':
|
||||||
|
if scanFollowsMultilineLiteralStringDelimiter(b[i:]) {
|
||||||
|
i += 3
|
||||||
|
|
||||||
|
// At that point we found 3 apostrophe, and i is the
|
||||||
|
// index of the byte after the third one. The scanner
|
||||||
|
// needs to be eager, because there can be an extra 2
|
||||||
|
// apostrophe that can be accepted at the end of the
|
||||||
|
// string.
|
||||||
|
|
||||||
|
if i >= len(b) || b[i] != '\'' {
|
||||||
|
return b[:i], b[i:], nil
|
||||||
|
}
|
||||||
|
i++
|
||||||
|
|
||||||
|
if i >= len(b) || b[i] != '\'' {
|
||||||
|
return b[:i], b[i:], nil
|
||||||
|
}
|
||||||
|
i++
|
||||||
|
|
||||||
|
if i < len(b) && b[i] == '\'' {
|
||||||
|
return nil, b[i:], newDecodeError(b[i-3:i+1], "''' not allowed in multiline literal string")
|
||||||
|
}
|
||||||
|
|
||||||
|
return b[:i], b[i:], nil
|
||||||
|
}
|
||||||
|
case '\r':
|
||||||
|
if len(b) < i+2 {
|
||||||
|
return nil, b[i:], newDecodeError(b[len(b):], `need a \n after \r`)
|
||||||
|
}
|
||||||
|
if b[i+1] != '\n' {
|
||||||
|
return nil, b[i+2:], newDecodeError(b[i:i+2], `need a \n after \r`)
|
||||||
|
}
|
||||||
|
i += 2 // skip the \n
|
||||||
|
continue
|
||||||
}
|
}
|
||||||
|
size := utf8ValidNext(b[i:])
|
||||||
|
if size == 0 {
|
||||||
|
return nil, b[i:], newDecodeError(b[i:i+1], "invalid character")
|
||||||
|
}
|
||||||
|
i += size
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil, nil, newDecodeError(b[len(b):], `multiline literal string not terminated by '''`)
|
return nil, b[len(b):], newDecodeError(b[len(b):], `multiline literal string not terminated by '''`)
|
||||||
}
|
}
|
||||||
|
|
||||||
func scanWindowsNewline(b []byte) ([]byte, []byte, error) {
|
func scanWindowsNewline(b []byte) ([]byte, []byte, error) {
|
||||||
const lenCRLF = 2
|
const lenCRLF = 2
|
||||||
if len(b) < lenCRLF {
|
if len(b) < lenCRLF {
|
||||||
return nil, nil, newDecodeError(b, "windows new line expected")
|
return nil, b, newDecodeError(b, "windows new line expected")
|
||||||
}
|
}
|
||||||
|
|
||||||
if b[1] != '\n' {
|
if b[1] != '\n' {
|
||||||
return nil, nil, newDecodeError(b, `windows new line should be \r\n`)
|
return nil, b[2:], newDecodeError(b, `windows new line should be \r\n`)
|
||||||
}
|
}
|
||||||
|
|
||||||
return b[:lenCRLF], b[lenCRLF:], nil
|
return b[:lenCRLF], b[lenCRLF:], nil
|
||||||
@@ -106,45 +150,62 @@ func scanWhitespace(b []byte) ([]byte, []byte) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
//nolint:unparam
|
//nolint:unparam
|
||||||
func scanComment(b []byte) ([]byte, []byte) {
|
func scanComment(b []byte) ([]byte, []byte, error) {
|
||||||
// comment-start-symbol = %x23 ; #
|
// comment-start-symbol = %x23 ; #
|
||||||
// non-ascii = %x80-D7FF / %xE000-10FFFF
|
// non-ascii = %x80-D7FF / %xE000-10FFFF
|
||||||
// non-eol = %x09 / %x20-7F / non-ascii
|
// non-eol = %x09 / %x20-7F / non-ascii
|
||||||
//
|
//
|
||||||
// comment = comment-start-symbol *non-eol
|
// comment = comment-start-symbol *non-eol
|
||||||
for i := 1; i < len(b); i++ {
|
|
||||||
|
for i := 1; i < len(b); {
|
||||||
if b[i] == '\n' {
|
if b[i] == '\n' {
|
||||||
return b[:i], b[i:]
|
return b[:i], b[i:], nil
|
||||||
}
|
}
|
||||||
|
if b[i] == '\r' {
|
||||||
|
if i+1 < len(b) && b[i+1] == '\n' {
|
||||||
|
return b[:i+1], b[i+1:], nil
|
||||||
|
}
|
||||||
|
return nil, nil, newDecodeError(b[i:i+1], "invalid character in comment")
|
||||||
|
}
|
||||||
|
size := utf8ValidNext(b[i:])
|
||||||
|
if size == 0 {
|
||||||
|
return nil, b[i+1:], newDecodeError(b[i:i+1], "invalid character in comment")
|
||||||
|
}
|
||||||
|
|
||||||
|
i += size
|
||||||
}
|
}
|
||||||
|
|
||||||
return b, nil
|
return b, b[len(b):], nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func scanBasicString(b []byte) ([]byte, []byte, error) {
|
func scanBasicString(b []byte) ([]byte, bool, []byte, error) {
|
||||||
// basic-string = quotation-mark *basic-char quotation-mark
|
// basic-string = quotation-mark *basic-char quotation-mark
|
||||||
// quotation-mark = %x22 ; "
|
// quotation-mark = %x22 ; "
|
||||||
// basic-char = basic-unescaped / escaped
|
// basic-char = basic-unescaped / escaped
|
||||||
// basic-unescaped = wschar / %x21 / %x23-5B / %x5D-7E / non-ascii
|
// basic-unescaped = wschar / %x21 / %x23-5B / %x5D-7E / non-ascii
|
||||||
// escaped = escape escape-seq-char
|
// escaped = escape escape-seq-char
|
||||||
for i := 1; i < len(b); i++ {
|
escaped := false
|
||||||
|
i := 1
|
||||||
|
|
||||||
|
for ; i < len(b); i++ {
|
||||||
switch b[i] {
|
switch b[i] {
|
||||||
case '"':
|
case '"':
|
||||||
return b[:i+1], b[i+1:], nil
|
return b[:i+1], escaped, b[i+1:], nil
|
||||||
case '\n':
|
case '\n', '\r':
|
||||||
return nil, nil, newDecodeError(b[i:i+1], "basic strings cannot have new lines")
|
return nil, escaped, b[i+1:], newDecodeError(b[i:i+1], "basic strings cannot have new lines")
|
||||||
case '\\':
|
case '\\':
|
||||||
if len(b) < i+2 {
|
if len(b) < i+2 {
|
||||||
return nil, nil, newDecodeError(b[i:i+1], "need a character after \\")
|
return nil, escaped, b[i+1:], newDecodeError(b[i:i+1], "need a character after \\")
|
||||||
}
|
}
|
||||||
|
escaped = true
|
||||||
i++ // skip the next character
|
i++ // skip the next character
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil, nil, newDecodeError(b[len(b):], `basic string not terminated by "`)
|
return nil, escaped, b[len(b):], newDecodeError(b[len(b):], `basic string not terminated by "`)
|
||||||
}
|
}
|
||||||
|
|
||||||
func scanMultilineBasicString(b []byte) ([]byte, []byte, error) {
|
func scanMultilineBasicString(b []byte) ([]byte, bool, []byte, error) {
|
||||||
// ml-basic-string = ml-basic-string-delim [ newline ] ml-basic-body
|
// ml-basic-string = ml-basic-string-delim [ newline ] ml-basic-body
|
||||||
// ml-basic-string-delim
|
// ml-basic-string-delim
|
||||||
// ml-basic-string-delim = 3quotation-mark
|
// ml-basic-string-delim = 3quotation-mark
|
||||||
@@ -155,19 +216,54 @@ func scanMultilineBasicString(b []byte) ([]byte, []byte, error) {
|
|||||||
// mlb-quotes = 1*2quotation-mark
|
// mlb-quotes = 1*2quotation-mark
|
||||||
// mlb-unescaped = wschar / %x21 / %x23-5B / %x5D-7E / non-ascii
|
// mlb-unescaped = wschar / %x21 / %x23-5B / %x5D-7E / non-ascii
|
||||||
// mlb-escaped-nl = escape ws newline *( wschar / newline )
|
// mlb-escaped-nl = escape ws newline *( wschar / newline )
|
||||||
for i := 3; i < len(b); i++ {
|
|
||||||
|
escaped := false
|
||||||
|
i := 3
|
||||||
|
|
||||||
|
for ; i < len(b); i++ {
|
||||||
switch b[i] {
|
switch b[i] {
|
||||||
case '"':
|
case '"':
|
||||||
if scanFollowsMultilineBasicStringDelimiter(b[i:]) {
|
if scanFollowsMultilineBasicStringDelimiter(b[i:]) {
|
||||||
return b[:i+3], b[i+3:], nil
|
i += 3
|
||||||
|
|
||||||
|
// At that point we found 3 apostrophe, and i is the
|
||||||
|
// index of the byte after the third one. The scanner
|
||||||
|
// needs to be eager, because there can be an extra 2
|
||||||
|
// apostrophe that can be accepted at the end of the
|
||||||
|
// string.
|
||||||
|
|
||||||
|
if i >= len(b) || b[i] != '"' {
|
||||||
|
return b[:i], escaped, b[i:], nil
|
||||||
|
}
|
||||||
|
i++
|
||||||
|
|
||||||
|
if i >= len(b) || b[i] != '"' {
|
||||||
|
return b[:i], escaped, b[i:], nil
|
||||||
|
}
|
||||||
|
i++
|
||||||
|
|
||||||
|
if i < len(b) && b[i] == '"' {
|
||||||
|
return nil, escaped, b[i+1:], newDecodeError(b[i-3:i+1], `""" not allowed in multiline basic string`)
|
||||||
|
}
|
||||||
|
|
||||||
|
return b[:i], escaped, b[i:], nil
|
||||||
}
|
}
|
||||||
case '\\':
|
case '\\':
|
||||||
if len(b) < i+2 {
|
if len(b) < i+2 {
|
||||||
return nil, nil, newDecodeError(b[len(b):], "need a character after \\")
|
return nil, escaped, b[len(b):], newDecodeError(b[len(b):], "need a character after \\")
|
||||||
}
|
}
|
||||||
|
escaped = true
|
||||||
i++ // skip the next character
|
i++ // skip the next character
|
||||||
|
case '\r':
|
||||||
|
if len(b) < i+2 {
|
||||||
|
return nil, escaped, b[len(b):], newDecodeError(b[len(b):], `need a \n after \r`)
|
||||||
|
}
|
||||||
|
if b[i+1] != '\n' {
|
||||||
|
return nil, escaped, b[i+2:], newDecodeError(b[i:i+2], `need a \n after \r`)
|
||||||
|
}
|
||||||
|
i++ // skip the \n
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil, nil, newDecodeError(b[len(b):], `multiline basic string not terminated by """`)
|
return nil, escaped, b[len(b):], newDecodeError(b[len(b):], `multiline basic string not terminated by """`)
|
||||||
}
|
}
|
||||||
|
|||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("0=0000-01-01 00:00:00")
|
||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("\"\\n\"=\"\"")
|
||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("''=0")
|
||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("0=0000-01-01")
|
||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("0=\"\"\"\\U00000000\"\"\"")
|
||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("0=[[{}]]")
|
||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("\"\\b\"=\"\"")
|
||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("0=inf")
|
||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("0=0000-01-01 00:00:00+00:00")
|
||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("0=[{}]")
|
||||||
Vendored
+2
@@ -0,0 +1,2 @@
|
|||||||
|
go test fuzz v1
|
||||||
|
[]byte("0=nan")
|
||||||
+23
-117
@@ -1,14 +1,14 @@
|
|||||||
|
//go:generate go run ./cmd/tomltestgen/main.go -o toml_testgen_test.go
|
||||||
|
|
||||||
// This is a support file for toml_testgen_test.go
|
// This is a support file for toml_testgen_test.go
|
||||||
package toml_test
|
package toml_test
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
|
||||||
"strconv"
|
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
|
||||||
|
|
||||||
"github.com/pelletier/go-toml/v2"
|
"github.com/pelletier/go-toml/v2"
|
||||||
|
"github.com/pelletier/go-toml/v2/internal/testsuite"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -17,10 +17,14 @@ func testgenInvalid(t *testing.T, input string) {
|
|||||||
t.Logf("Input TOML:\n%s", input)
|
t.Logf("Input TOML:\n%s", input)
|
||||||
|
|
||||||
doc := map[string]interface{}{}
|
doc := map[string]interface{}{}
|
||||||
err := toml.Unmarshal([]byte(input), &doc)
|
err := testsuite.Unmarshal([]byte(input), &doc)
|
||||||
|
|
||||||
if err == nil {
|
if err == nil {
|
||||||
t.Log(json.Marshal(doc))
|
out, err := json.Marshal(doc)
|
||||||
|
if err != nil {
|
||||||
|
panic("could not marshal map to json")
|
||||||
|
}
|
||||||
|
t.Log("JSON output from unmarshal:", string(out))
|
||||||
t.Fatalf("test did not fail")
|
t.Fatalf("test did not fail")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -29,124 +33,26 @@ func testgenValid(t *testing.T, input string, jsonRef string) {
|
|||||||
t.Helper()
|
t.Helper()
|
||||||
t.Logf("Input TOML:\n%s", input)
|
t.Logf("Input TOML:\n%s", input)
|
||||||
|
|
||||||
doc := map[string]interface{}{}
|
// TODO: change this to interface{}
|
||||||
|
var doc map[string]interface{}
|
||||||
|
|
||||||
err := toml.Unmarshal([]byte(input), &doc)
|
err := testsuite.Unmarshal([]byte(input), &doc)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
if de, ok := err.(*toml.DecodeError); ok {
|
||||||
|
t.Logf("%s\n%s", err, de)
|
||||||
|
}
|
||||||
t.Fatalf("failed parsing toml: %s", err)
|
t.Fatalf("failed parsing toml: %s", err)
|
||||||
}
|
}
|
||||||
|
j, err := testsuite.ValueToTaggedJSON(doc)
|
||||||
refDoc := testgenBuildRefDoc(jsonRef)
|
|
||||||
|
|
||||||
require.Equal(t, refDoc, doc)
|
|
||||||
|
|
||||||
out, err := toml.Marshal(doc)
|
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
doc2 := map[string]interface{}{}
|
var ref interface{}
|
||||||
err = toml.Unmarshal(out, &doc2)
|
err = json.Unmarshal([]byte(jsonRef), &ref)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
require.Equal(t, refDoc, doc2)
|
var actual interface{}
|
||||||
}
|
err = json.Unmarshal([]byte(j), &actual)
|
||||||
|
require.NoError(t, err)
|
||||||
func testgenBuildRefDoc(jsonRef string) map[string]interface{} {
|
|
||||||
descTree := map[string]interface{}{}
|
testsuite.CmpJSON(t, "", ref, actual)
|
||||||
|
|
||||||
err := json.Unmarshal([]byte(jsonRef), &descTree)
|
|
||||||
if err != nil {
|
|
||||||
panic(fmt.Sprintf("reference doc should be valid JSON: %s", err))
|
|
||||||
}
|
|
||||||
|
|
||||||
doc := testGenTranslateDesc(descTree)
|
|
||||||
if doc == nil {
|
|
||||||
return map[string]interface{}{}
|
|
||||||
}
|
|
||||||
|
|
||||||
return doc.(map[string]interface{})
|
|
||||||
}
|
|
||||||
|
|
||||||
//nolint:funlen,gocognit,cyclop
|
|
||||||
func testGenTranslateDesc(input interface{}) interface{} {
|
|
||||||
a, ok := input.([]interface{})
|
|
||||||
if ok {
|
|
||||||
xs := make([]interface{}, len(a))
|
|
||||||
for i, v := range a {
|
|
||||||
xs[i] = testGenTranslateDesc(v)
|
|
||||||
}
|
|
||||||
|
|
||||||
return xs
|
|
||||||
}
|
|
||||||
|
|
||||||
d, ok := input.(map[string]interface{})
|
|
||||||
if !ok {
|
|
||||||
panic(fmt.Sprintf("input should be valid map[string]: %v", input))
|
|
||||||
}
|
|
||||||
|
|
||||||
var (
|
|
||||||
dtype string
|
|
||||||
dvalue interface{}
|
|
||||||
)
|
|
||||||
|
|
||||||
//nolint:nestif
|
|
||||||
if len(d) == 2 {
|
|
||||||
dtypeiface, ok := d["type"]
|
|
||||||
if ok {
|
|
||||||
dvalue, ok = d["value"]
|
|
||||||
if ok {
|
|
||||||
dtype = dtypeiface.(string)
|
|
||||||
|
|
||||||
switch dtype {
|
|
||||||
case "string":
|
|
||||||
return dvalue.(string)
|
|
||||||
case "float":
|
|
||||||
v, err := strconv.ParseFloat(dvalue.(string), 64)
|
|
||||||
if err != nil {
|
|
||||||
panic(fmt.Sprintf("invalid float '%s': %s", dvalue, err))
|
|
||||||
}
|
|
||||||
|
|
||||||
return v
|
|
||||||
case "integer":
|
|
||||||
v, err := strconv.ParseInt(dvalue.(string), 10, 64)
|
|
||||||
if err != nil {
|
|
||||||
panic(fmt.Sprintf("invalid int '%s': %s", dvalue, err))
|
|
||||||
}
|
|
||||||
|
|
||||||
return v
|
|
||||||
case "bool":
|
|
||||||
return dvalue.(string) == "true"
|
|
||||||
case "datetime":
|
|
||||||
dt, err := time.Parse("2006-01-02T15:04:05Z", dvalue.(string))
|
|
||||||
if err != nil {
|
|
||||||
panic(fmt.Sprintf("invalid datetime '%s': %s", dvalue, err))
|
|
||||||
}
|
|
||||||
|
|
||||||
return dt
|
|
||||||
case "array":
|
|
||||||
if dvalue == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
a := dvalue.([]interface{})
|
|
||||||
|
|
||||||
xs := make([]interface{}, len(a))
|
|
||||||
|
|
||||||
for i, v := range a {
|
|
||||||
xs[i] = testGenTranslateDesc(v)
|
|
||||||
}
|
|
||||||
|
|
||||||
return xs
|
|
||||||
}
|
|
||||||
|
|
||||||
panic(fmt.Sprintf("unknown type: %s", dtype))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
dest := map[string]interface{}{}
|
|
||||||
for k, v := range d {
|
|
||||||
dest[k] = testGenTranslateDesc(v)
|
|
||||||
}
|
|
||||||
|
|
||||||
return dest
|
|
||||||
}
|
}
|
||||||
|
|||||||
+1486
-772
File diff suppressed because it is too large
Load Diff
@@ -11,3 +11,4 @@ var textMarshalerType = reflect.TypeOf(new(encoding.TextMarshaler)).Elem()
|
|||||||
var textUnmarshalerType = reflect.TypeOf(new(encoding.TextUnmarshaler)).Elem()
|
var textUnmarshalerType = reflect.TypeOf(new(encoding.TextUnmarshaler)).Elem()
|
||||||
var mapStringInterfaceType = reflect.TypeOf(map[string]interface{}{})
|
var mapStringInterfaceType = reflect.TypeOf(map[string]interface{}{})
|
||||||
var sliceInterfaceType = reflect.TypeOf([]interface{}{})
|
var sliceInterfaceType = reflect.TypeOf([]interface{}{})
|
||||||
|
var stringType = reflect.TypeOf("")
|
||||||
|
|||||||
+270
-142
@@ -9,10 +9,11 @@ import (
|
|||||||
"math"
|
"math"
|
||||||
"reflect"
|
"reflect"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/pelletier/go-toml/v2/internal/ast"
|
"github.com/pelletier/go-toml/v2/internal/ast"
|
||||||
|
"github.com/pelletier/go-toml/v2/internal/danger"
|
||||||
"github.com/pelletier/go-toml/v2/internal/tracker"
|
"github.com/pelletier/go-toml/v2/internal/tracker"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -41,24 +42,27 @@ func NewDecoder(r io.Reader) *Decoder {
|
|||||||
return &Decoder{r: r}
|
return &Decoder{r: r}
|
||||||
}
|
}
|
||||||
|
|
||||||
// SetStrict toggles decoding in stict mode.
|
// DisallowUnknownFields causes the Decoder to return an error when the
|
||||||
|
// destination is a struct and the input contains a key that does not match a
|
||||||
|
// non-ignored field.
|
||||||
//
|
//
|
||||||
// When the decoder is in strict mode, it will record fields from the document
|
// In that case, the Decoder returns a StrictMissingError that can be used to
|
||||||
// that could not be set on the target value. In that case, the decoder returns
|
// retrieve the individual errors as well as generate a human readable
|
||||||
// a StrictMissingError that can be used to retrieve the individual errors as
|
// description of the missing fields.
|
||||||
// well as generate a human readable description of the missing fields.
|
func (d *Decoder) DisallowUnknownFields() *Decoder {
|
||||||
func (d *Decoder) SetStrict(strict bool) {
|
d.strict = true
|
||||||
d.strict = strict
|
return d
|
||||||
}
|
}
|
||||||
|
|
||||||
// Decode the whole content of r into v.
|
// Decode the whole content of r into v.
|
||||||
//
|
//
|
||||||
// By default, values in the document that don't exist in the target Go value
|
// By default, values in the document that don't exist in the target Go value
|
||||||
// are ignored. See Decoder.SetStrict() to change this behavior.
|
// are ignored. See Decoder.DisallowUnknownFields() to change this behavior.
|
||||||
//
|
//
|
||||||
// When a TOML local date, time, or date-time is decoded into a time.Time, its
|
// When a TOML local date, time, or date-time is decoded into a time.Time, its
|
||||||
// value is represented in time.Local timezone. Otherwise the approriate Local*
|
// value is represented in time.Local timezone. Otherwise the approriate Local*
|
||||||
// structure is used.
|
// structure is used. For time values, precision up to the nanosecond is
|
||||||
|
// supported by truncating extra digits.
|
||||||
//
|
//
|
||||||
// Empty tables decoded in an interface{} create an empty initialized
|
// Empty tables decoded in an interface{} create an empty initialized
|
||||||
// map[string]interface{}.
|
// map[string]interface{}.
|
||||||
@@ -70,22 +74,27 @@ func (d *Decoder) SetStrict(strict bool) {
|
|||||||
// bounds for the target type (which includes negative numbers when decoding
|
// bounds for the target type (which includes negative numbers when decoding
|
||||||
// into an unsigned int).
|
// into an unsigned int).
|
||||||
//
|
//
|
||||||
// Type mapping
|
// If an error occurs while decoding the content of the document, this function
|
||||||
|
// returns a toml.DecodeError, providing context about the issue. When using
|
||||||
|
// strict mode and a field is missing, a `toml.StrictMissingError` is
|
||||||
|
// returned. In any other case, this function returns a standard Go error.
|
||||||
|
//
|
||||||
|
// # Type mapping
|
||||||
//
|
//
|
||||||
// List of supported TOML types and their associated accepted Go types:
|
// List of supported TOML types and their associated accepted Go types:
|
||||||
//
|
//
|
||||||
// String -> string
|
// String -> string
|
||||||
// Integer -> uint*, int*, depending on size
|
// Integer -> uint*, int*, depending on size
|
||||||
// Float -> float*, depending on size
|
// Float -> float*, depending on size
|
||||||
// Boolean -> bool
|
// Boolean -> bool
|
||||||
// Offset Date-Time -> time.Time
|
// Offset Date-Time -> time.Time
|
||||||
// Local Date-time -> LocalDateTime, time.Time
|
// Local Date-time -> LocalDateTime, time.Time
|
||||||
// Local Date -> LocalDate, time.Time
|
// Local Date -> LocalDate, time.Time
|
||||||
// Local Time -> LocalTime, time.Time
|
// Local Time -> LocalTime, time.Time
|
||||||
// Array -> slice and array, depending on elements types
|
// Array -> slice and array, depending on elements types
|
||||||
// Table -> map and struct
|
// Table -> map and struct
|
||||||
// Inline Table -> same as Table
|
// Inline Table -> same as Table
|
||||||
// Array of Tables -> same as Array and Table
|
// Array of Tables -> same as Array and Table
|
||||||
func (d *Decoder) Decode(v interface{}) error {
|
func (d *Decoder) Decode(v interface{}) error {
|
||||||
b, err := ioutil.ReadAll(d.r)
|
b, err := ioutil.ReadAll(d.r)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -114,7 +123,7 @@ type decoder struct {
|
|||||||
stashedExpr bool
|
stashedExpr bool
|
||||||
|
|
||||||
// Skip expressions until a table is found. This is set to true when a
|
// Skip expressions until a table is found. This is set to true when a
|
||||||
// table could not be create (missing field in map), so all KV expressions
|
// table could not be created (missing field in map), so all KV expressions
|
||||||
// need to be skipped.
|
// need to be skipped.
|
||||||
skipUntilTable bool
|
skipUntilTable bool
|
||||||
|
|
||||||
@@ -129,6 +138,23 @@ type decoder struct {
|
|||||||
|
|
||||||
// Strict mode
|
// Strict mode
|
||||||
strict strict
|
strict strict
|
||||||
|
|
||||||
|
// Current context for the error.
|
||||||
|
errorContext *errorContext
|
||||||
|
}
|
||||||
|
|
||||||
|
type errorContext struct {
|
||||||
|
Struct reflect.Type
|
||||||
|
Field []int
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) typeMismatchError(toml string, target reflect.Type) error {
|
||||||
|
if d.errorContext != nil && d.errorContext.Struct != nil {
|
||||||
|
ctx := d.errorContext
|
||||||
|
f := ctx.Struct.FieldByIndex(ctx.Field)
|
||||||
|
return fmt.Errorf("toml: cannot decode TOML %s into struct field %s.%s of type %s", toml, ctx.Struct, f.Name, f.Type)
|
||||||
|
}
|
||||||
|
return fmt.Errorf("toml: cannot decode TOML %s into a Go value of type %s", toml, target)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *decoder) expr() *ast.Node {
|
func (d *decoder) expr() *ast.Node {
|
||||||
@@ -174,7 +200,13 @@ func (d *decoder) FromParser(v interface{}) error {
|
|||||||
return fmt.Errorf("toml: decoding pointer target cannot be nil")
|
return fmt.Errorf("toml: decoding pointer target cannot be nil")
|
||||||
}
|
}
|
||||||
|
|
||||||
err := d.fromParser(r.Elem())
|
r = r.Elem()
|
||||||
|
if r.Kind() == reflect.Interface && r.IsNil() {
|
||||||
|
newMap := map[string]interface{}{}
|
||||||
|
r.Set(reflect.ValueOf(newMap))
|
||||||
|
}
|
||||||
|
|
||||||
|
err := d.fromParser(r)
|
||||||
if err == nil {
|
if err == nil {
|
||||||
return d.strict.Error(d.p.data)
|
return d.strict.Error(d.p.data)
|
||||||
}
|
}
|
||||||
@@ -290,10 +322,12 @@ func (d *decoder) handleArrayTableCollectionLast(key ast.Iterator, v reflect.Val
|
|||||||
return v, nil
|
return v, nil
|
||||||
case reflect.Slice:
|
case reflect.Slice:
|
||||||
elemType := v.Type().Elem()
|
elemType := v.Type().Elem()
|
||||||
|
var elem reflect.Value
|
||||||
if elemType.Kind() == reflect.Interface {
|
if elemType.Kind() == reflect.Interface {
|
||||||
elemType = mapStringInterfaceType
|
elem = makeMapStringInterface()
|
||||||
|
} else {
|
||||||
|
elem = reflect.New(elemType).Elem()
|
||||||
}
|
}
|
||||||
elem := reflect.New(elemType).Elem()
|
|
||||||
elem2, err := d.handleArrayTable(key, elem)
|
elem2, err := d.handleArrayTable(key, elem)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return reflect.Value{}, err
|
return reflect.Value{}, err
|
||||||
@@ -310,9 +344,9 @@ func (d *decoder) handleArrayTableCollectionLast(key ast.Iterator, v reflect.Val
|
|||||||
elem := v.Index(idx)
|
elem := v.Index(idx)
|
||||||
_, err := d.handleArrayTable(key, elem)
|
_, err := d.handleArrayTable(key, elem)
|
||||||
return v, err
|
return v, err
|
||||||
|
default:
|
||||||
|
return reflect.Value{}, fmt.Errorf("toml: cannot decode array table into a %s", v.Type())
|
||||||
}
|
}
|
||||||
|
|
||||||
return d.handleArrayTable(key, v)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// When parsing an array table expression, each part of the key needs to be
|
// When parsing an array table expression, each part of the key needs to be
|
||||||
@@ -337,7 +371,9 @@ func (d *decoder) handleArrayTableCollection(key ast.Iterator, v reflect.Value)
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return reflect.Value{}, err
|
return reflect.Value{}, err
|
||||||
}
|
}
|
||||||
v.Elem().Set(elem)
|
if elem.IsValid() {
|
||||||
|
v.Elem().Set(elem)
|
||||||
|
}
|
||||||
|
|
||||||
return v, nil
|
return v, nil
|
||||||
case reflect.Slice:
|
case reflect.Slice:
|
||||||
@@ -370,13 +406,23 @@ func (d *decoder) handleKeyPart(key ast.Iterator, v reflect.Value, nextFn handle
|
|||||||
// First, dispatch over v to make sure it is a valid object.
|
// First, dispatch over v to make sure it is a valid object.
|
||||||
// There is no guarantee over what it could be.
|
// There is no guarantee over what it could be.
|
||||||
switch v.Kind() {
|
switch v.Kind() {
|
||||||
|
case reflect.Ptr:
|
||||||
|
elem := v.Elem()
|
||||||
|
if !elem.IsValid() {
|
||||||
|
v.Set(reflect.New(v.Type().Elem()))
|
||||||
|
}
|
||||||
|
elem = v.Elem()
|
||||||
|
return d.handleKeyPart(key, elem, nextFn, makeFn)
|
||||||
case reflect.Map:
|
case reflect.Map:
|
||||||
// Create the key for the map element. For now assume it's a string.
|
vt := v.Type()
|
||||||
mk := reflect.ValueOf(string(key.Node().Data))
|
|
||||||
|
// Create the key for the map element. Convert to key type.
|
||||||
|
mk := reflect.ValueOf(string(key.Node().Data)).Convert(vt.Key())
|
||||||
|
|
||||||
// If the map does not exist, create it.
|
// If the map does not exist, create it.
|
||||||
if v.IsNil() {
|
if v.IsNil() {
|
||||||
v = reflect.MakeMap(v.Type())
|
vt := v.Type()
|
||||||
|
v = reflect.MakeMap(vt)
|
||||||
rv = v
|
rv = v
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -388,7 +434,7 @@ func (d *decoder) handleKeyPart(key ast.Iterator, v reflect.Value, nextFn handle
|
|||||||
// map[string]interface{} or a []interface{} depending on whether
|
// map[string]interface{} or a []interface{} depending on whether
|
||||||
// this is the last part of the array table key.
|
// this is the last part of the array table key.
|
||||||
|
|
||||||
t := v.Type().Elem()
|
t := vt.Elem()
|
||||||
if t.Kind() == reflect.Interface {
|
if t.Kind() == reflect.Interface {
|
||||||
mv = makeFn()
|
mv = makeFn()
|
||||||
} else {
|
} else {
|
||||||
@@ -401,6 +447,13 @@ func (d *decoder) handleKeyPart(key ast.Iterator, v reflect.Value, nextFn handle
|
|||||||
mv = makeFn()
|
mv = makeFn()
|
||||||
}
|
}
|
||||||
set = true
|
set = true
|
||||||
|
} else if !mv.CanAddr() {
|
||||||
|
vt := v.Type()
|
||||||
|
t := vt.Elem()
|
||||||
|
oldmv := mv
|
||||||
|
mv = reflect.New(t).Elem()
|
||||||
|
mv.Set(oldmv)
|
||||||
|
set = true
|
||||||
}
|
}
|
||||||
|
|
||||||
x, err := nextFn(key, mv)
|
x, err := nextFn(key, mv)
|
||||||
@@ -417,12 +470,20 @@ func (d *decoder) handleKeyPart(key ast.Iterator, v reflect.Value, nextFn handle
|
|||||||
v.SetMapIndex(mk, mv)
|
v.SetMapIndex(mk, mv)
|
||||||
}
|
}
|
||||||
case reflect.Struct:
|
case reflect.Struct:
|
||||||
f, found := structField(v, string(key.Node().Data))
|
path, found := structFieldPath(v, string(key.Node().Data))
|
||||||
if !found {
|
if !found {
|
||||||
d.skipUntilTable = true
|
d.skipUntilTable = true
|
||||||
return reflect.Value{}, nil
|
return reflect.Value{}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if d.errorContext == nil {
|
||||||
|
d.errorContext = new(errorContext)
|
||||||
|
}
|
||||||
|
t := v.Type()
|
||||||
|
d.errorContext.Struct = t
|
||||||
|
d.errorContext.Field = path
|
||||||
|
|
||||||
|
f := fieldByIndex(v, path)
|
||||||
x, err := nextFn(key, f)
|
x, err := nextFn(key, f)
|
||||||
if err != nil || d.skipUntilTable {
|
if err != nil || d.skipUntilTable {
|
||||||
return reflect.Value{}, err
|
return reflect.Value{}, err
|
||||||
@@ -430,11 +491,13 @@ func (d *decoder) handleKeyPart(key ast.Iterator, v reflect.Value, nextFn handle
|
|||||||
if x.IsValid() {
|
if x.IsValid() {
|
||||||
f.Set(x)
|
f.Set(x)
|
||||||
}
|
}
|
||||||
|
d.errorContext.Field = nil
|
||||||
|
d.errorContext.Struct = nil
|
||||||
case reflect.Interface:
|
case reflect.Interface:
|
||||||
if v.Elem().IsValid() {
|
if v.Elem().IsValid() {
|
||||||
v = v.Elem()
|
v = v.Elem()
|
||||||
} else {
|
} else {
|
||||||
v = reflect.MakeMap(mapStringInterfaceType)
|
v = makeMapStringInterface()
|
||||||
}
|
}
|
||||||
|
|
||||||
x, err := d.handleKeyPart(key, v, nextFn, makeFn)
|
x, err := d.handleKeyPart(key, v, nextFn, makeFn)
|
||||||
@@ -469,6 +532,9 @@ func (d *decoder) handleArrayTablePart(key ast.Iterator, v reflect.Value) (refle
|
|||||||
// cannot handle it.
|
// cannot handle it.
|
||||||
func (d *decoder) handleTable(key ast.Iterator, v reflect.Value) (reflect.Value, error) {
|
func (d *decoder) handleTable(key ast.Iterator, v reflect.Value) (reflect.Value, error) {
|
||||||
if v.Kind() == reflect.Slice {
|
if v.Kind() == reflect.Slice {
|
||||||
|
if v.Len() == 0 {
|
||||||
|
return reflect.Value{}, newDecodeError(key.Node().Data, "cannot store a table in a slice")
|
||||||
|
}
|
||||||
elem := v.Index(v.Len() - 1)
|
elem := v.Index(v.Len() - 1)
|
||||||
x, err := d.handleTable(key, elem)
|
x, err := d.handleTable(key, elem)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -503,6 +569,11 @@ func (d *decoder) handleKeyValues(v reflect.Value) (reflect.Value, error) {
|
|||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
|
err := d.seen.CheckExpression(expr)
|
||||||
|
if err != nil {
|
||||||
|
return reflect.Value{}, err
|
||||||
|
}
|
||||||
|
|
||||||
x, err := d.handleKeyValue(expr, v)
|
x, err := d.handleKeyValue(expr, v)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return reflect.Value{}, err
|
return reflect.Value{}, err
|
||||||
@@ -533,10 +604,6 @@ func (d *decoder) handleTablePart(key ast.Iterator, v reflect.Value) (reflect.Va
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (d *decoder) tryTextUnmarshaler(node *ast.Node, v reflect.Value) (bool, error) {
|
func (d *decoder) tryTextUnmarshaler(node *ast.Node, v reflect.Value) (bool, error) {
|
||||||
if v.Kind() != reflect.Struct {
|
|
||||||
return false, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Special case for time, because we allow to unmarshal to it from
|
// Special case for time, because we allow to unmarshal to it from
|
||||||
// different kind of AST nodes.
|
// different kind of AST nodes.
|
||||||
if v.Type() == timeType {
|
if v.Type() == timeType {
|
||||||
@@ -546,7 +613,7 @@ func (d *decoder) tryTextUnmarshaler(node *ast.Node, v reflect.Value) (bool, err
|
|||||||
if v.CanAddr() && v.Addr().Type().Implements(textUnmarshalerType) {
|
if v.CanAddr() && v.Addr().Type().Implements(textUnmarshalerType) {
|
||||||
err := v.Addr().Interface().(encoding.TextUnmarshaler).UnmarshalText(node.Data)
|
err := v.Addr().Interface().(encoding.TextUnmarshaler).UnmarshalText(node.Data)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return false, newDecodeError(d.p.Raw(node.Raw), "error calling UnmarshalText: %w", err)
|
return false, newDecodeError(d.p.Raw(node.Raw), "%w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
return true, nil
|
return true, nil
|
||||||
@@ -578,6 +645,8 @@ func (d *decoder) handleValue(value *ast.Node, v reflect.Value) error {
|
|||||||
return d.unmarshalDateTime(value, v)
|
return d.unmarshalDateTime(value, v)
|
||||||
case ast.LocalDate:
|
case ast.LocalDate:
|
||||||
return d.unmarshalLocalDate(value, v)
|
return d.unmarshalLocalDate(value, v)
|
||||||
|
case ast.LocalTime:
|
||||||
|
return d.unmarshalLocalTime(value, v)
|
||||||
case ast.LocalDateTime:
|
case ast.LocalDateTime:
|
||||||
return d.unmarshalLocalDateTime(value, v)
|
return d.unmarshalLocalDateTime(value, v)
|
||||||
case ast.InlineTable:
|
case ast.InlineTable:
|
||||||
@@ -624,7 +693,7 @@ func (d *decoder) unmarshalArray(array *ast.Node, v reflect.Value) error {
|
|||||||
default:
|
default:
|
||||||
// TODO: use newDecodeError, but first the parser needs to fill
|
// TODO: use newDecodeError, but first the parser needs to fill
|
||||||
// array.Data.
|
// array.Data.
|
||||||
return fmt.Errorf("toml: cannot store array in Go type %s", v.Kind())
|
return d.typeMismatchError("array", v.Type())
|
||||||
}
|
}
|
||||||
|
|
||||||
elemType := v.Type().Elem()
|
elemType := v.Type().Elem()
|
||||||
@@ -672,7 +741,7 @@ func (d *decoder) unmarshalInlineTable(itable *ast.Node, v reflect.Value) error
|
|||||||
case reflect.Interface:
|
case reflect.Interface:
|
||||||
elem := v.Elem()
|
elem := v.Elem()
|
||||||
if !elem.IsValid() {
|
if !elem.IsValid() {
|
||||||
elem = reflect.MakeMap(mapStringInterfaceType)
|
elem = makeMapStringInterface()
|
||||||
v.Set(elem)
|
v.Set(elem)
|
||||||
}
|
}
|
||||||
return d.unmarshalInlineTable(itable, elem)
|
return d.unmarshalInlineTable(itable, elem)
|
||||||
@@ -713,8 +782,7 @@ func (d *decoder) unmarshalLocalDate(value *ast.Node, v reflect.Value) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if v.Type() == timeType {
|
if v.Type() == timeType {
|
||||||
cast := ld.In(time.Local)
|
cast := ld.AsTime(time.Local)
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(cast))
|
v.Set(reflect.ValueOf(cast))
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@@ -724,6 +792,20 @@ func (d *decoder) unmarshalLocalDate(value *ast.Node, v reflect.Value) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (d *decoder) unmarshalLocalTime(value *ast.Node, v reflect.Value) error {
|
||||||
|
lt, rest, err := parseLocalTime(value.Data)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(rest) > 0 {
|
||||||
|
return newDecodeError(rest, "extra characters at the end of a local time")
|
||||||
|
}
|
||||||
|
|
||||||
|
v.Set(reflect.ValueOf(lt))
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func (d *decoder) unmarshalLocalDateTime(value *ast.Node, v reflect.Value) error {
|
func (d *decoder) unmarshalLocalDateTime(value *ast.Node, v reflect.Value) error {
|
||||||
ldt, rest, err := parseLocalDateTime(value.Data)
|
ldt, rest, err := parseLocalDateTime(value.Data)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -735,7 +817,7 @@ func (d *decoder) unmarshalLocalDateTime(value *ast.Node, v reflect.Value) error
|
|||||||
}
|
}
|
||||||
|
|
||||||
if v.Type() == timeType {
|
if v.Type() == timeType {
|
||||||
cast := ldt.In(time.Local)
|
cast := ldt.AsTime(time.Local)
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(cast))
|
v.Set(reflect.ValueOf(cast))
|
||||||
return nil
|
return nil
|
||||||
@@ -784,97 +866,118 @@ func (d *decoder) unmarshalFloat(value *ast.Node, v reflect.Value) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *decoder) unmarshalInteger(value *ast.Node, v reflect.Value) error {
|
const (
|
||||||
const (
|
maxInt = int64(^uint(0) >> 1)
|
||||||
maxInt = int64(^uint(0) >> 1)
|
minInt = -maxInt - 1
|
||||||
minInt = -maxInt - 1
|
)
|
||||||
)
|
|
||||||
|
|
||||||
|
// Maximum value of uint for decoding. Currently the decoder parses the integer
|
||||||
|
// into an int64. As a result, on architectures where uint is 64 bits, the
|
||||||
|
// effective maximum uint we can decode is the maximum of int64. On
|
||||||
|
// architectures where uint is 32 bits, the maximum value we can decode is
|
||||||
|
// lower: the maximum of uint32. I didn't find a way to figure out this value at
|
||||||
|
// compile time, so it is computed during initialization.
|
||||||
|
var maxUint int64 = math.MaxInt64
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
m := uint64(^uint(0))
|
||||||
|
if m < uint64(maxUint) {
|
||||||
|
maxUint = int64(m)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d *decoder) unmarshalInteger(value *ast.Node, v reflect.Value) error {
|
||||||
i, err := parseInteger(value.Data)
|
i, err := parseInteger(value.Data)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var r reflect.Value
|
||||||
|
|
||||||
switch v.Kind() {
|
switch v.Kind() {
|
||||||
case reflect.Int64:
|
case reflect.Int64:
|
||||||
v.SetInt(i)
|
v.SetInt(i)
|
||||||
|
return nil
|
||||||
case reflect.Int32:
|
case reflect.Int32:
|
||||||
if i < math.MinInt32 || i > math.MaxInt32 {
|
if i < math.MinInt32 || i > math.MaxInt32 {
|
||||||
return fmt.Errorf("toml: number %d does not fit in an int32", i)
|
return fmt.Errorf("toml: number %d does not fit in an int32", i)
|
||||||
}
|
}
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(int32(i)))
|
r = reflect.ValueOf(int32(i))
|
||||||
return nil
|
|
||||||
case reflect.Int16:
|
case reflect.Int16:
|
||||||
if i < math.MinInt16 || i > math.MaxInt16 {
|
if i < math.MinInt16 || i > math.MaxInt16 {
|
||||||
return fmt.Errorf("toml: number %d does not fit in an int16", i)
|
return fmt.Errorf("toml: number %d does not fit in an int16", i)
|
||||||
}
|
}
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(int16(i)))
|
r = reflect.ValueOf(int16(i))
|
||||||
case reflect.Int8:
|
case reflect.Int8:
|
||||||
if i < math.MinInt8 || i > math.MaxInt8 {
|
if i < math.MinInt8 || i > math.MaxInt8 {
|
||||||
return fmt.Errorf("toml: number %d does not fit in an int8", i)
|
return fmt.Errorf("toml: number %d does not fit in an int8", i)
|
||||||
}
|
}
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(int8(i)))
|
r = reflect.ValueOf(int8(i))
|
||||||
case reflect.Int:
|
case reflect.Int:
|
||||||
if i < minInt || i > maxInt {
|
if i < minInt || i > maxInt {
|
||||||
return fmt.Errorf("toml: number %d does not fit in an int", i)
|
return fmt.Errorf("toml: number %d does not fit in an int", i)
|
||||||
}
|
}
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(int(i)))
|
r = reflect.ValueOf(int(i))
|
||||||
case reflect.Uint64:
|
case reflect.Uint64:
|
||||||
if i < 0 {
|
if i < 0 {
|
||||||
return fmt.Errorf("toml: negative number %d does not fit in an uint64", i)
|
return fmt.Errorf("toml: negative number %d does not fit in an uint64", i)
|
||||||
}
|
}
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(uint64(i)))
|
r = reflect.ValueOf(uint64(i))
|
||||||
case reflect.Uint32:
|
case reflect.Uint32:
|
||||||
if i < 0 || i > math.MaxUint32 {
|
if i < 0 || i > math.MaxUint32 {
|
||||||
return fmt.Errorf("toml: negative number %d does not fit in an uint32", i)
|
return fmt.Errorf("toml: negative number %d does not fit in an uint32", i)
|
||||||
}
|
}
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(uint32(i)))
|
r = reflect.ValueOf(uint32(i))
|
||||||
case reflect.Uint16:
|
case reflect.Uint16:
|
||||||
if i < 0 || i > math.MaxUint16 {
|
if i < 0 || i > math.MaxUint16 {
|
||||||
return fmt.Errorf("toml: negative number %d does not fit in an uint16", i)
|
return fmt.Errorf("toml: negative number %d does not fit in an uint16", i)
|
||||||
}
|
}
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(uint16(i)))
|
r = reflect.ValueOf(uint16(i))
|
||||||
case reflect.Uint8:
|
case reflect.Uint8:
|
||||||
if i < 0 || i > math.MaxUint8 {
|
if i < 0 || i > math.MaxUint8 {
|
||||||
return fmt.Errorf("toml: negative number %d does not fit in an uint8", i)
|
return fmt.Errorf("toml: negative number %d does not fit in an uint8", i)
|
||||||
}
|
}
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(uint8(i)))
|
r = reflect.ValueOf(uint8(i))
|
||||||
case reflect.Uint:
|
case reflect.Uint:
|
||||||
if i < 0 {
|
if i < 0 || i > maxUint {
|
||||||
return fmt.Errorf("toml: negative number %d does not fit in an uint", i)
|
return fmt.Errorf("toml: negative number %d does not fit in an uint", i)
|
||||||
}
|
}
|
||||||
|
|
||||||
v.Set(reflect.ValueOf(uint(i)))
|
r = reflect.ValueOf(uint(i))
|
||||||
case reflect.Interface:
|
case reflect.Interface:
|
||||||
v.Set(reflect.ValueOf(i))
|
r = reflect.ValueOf(i)
|
||||||
default:
|
default:
|
||||||
err = fmt.Errorf("toml: cannot store TOML integer into a Go %s", v.Kind())
|
return d.typeMismatchError("integer", v.Type())
|
||||||
}
|
}
|
||||||
|
|
||||||
return err
|
if !r.Type().AssignableTo(v.Type()) {
|
||||||
|
r = r.Convert(v.Type())
|
||||||
|
}
|
||||||
|
|
||||||
|
v.Set(r)
|
||||||
|
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *decoder) unmarshalString(value *ast.Node, v reflect.Value) error {
|
func (d *decoder) unmarshalString(value *ast.Node, v reflect.Value) error {
|
||||||
var err error
|
|
||||||
|
|
||||||
switch v.Kind() {
|
switch v.Kind() {
|
||||||
case reflect.String:
|
case reflect.String:
|
||||||
v.SetString(string(value.Data))
|
v.SetString(string(value.Data))
|
||||||
case reflect.Interface:
|
case reflect.Interface:
|
||||||
v.Set(reflect.ValueOf(string(value.Data)))
|
v.Set(reflect.ValueOf(string(value.Data)))
|
||||||
default:
|
default:
|
||||||
err = newDecodeError(d.p.Raw(value.Raw), "cannot store TOML string into a Go %s", v.Kind())
|
return newDecodeError(d.p.Raw(value.Raw), "cannot store TOML string into a Go %s", v.Kind())
|
||||||
}
|
}
|
||||||
|
|
||||||
return err
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (d *decoder) handleKeyValue(expr *ast.Node, v reflect.Value) (reflect.Value, error) {
|
func (d *decoder) handleKeyValue(expr *ast.Node, v reflect.Value) (reflect.Value, error) {
|
||||||
@@ -909,12 +1012,15 @@ func (d *decoder) handleKeyValuePart(key ast.Iterator, value *ast.Node, v reflec
|
|||||||
// There is no guarantee over what it could be.
|
// There is no guarantee over what it could be.
|
||||||
switch v.Kind() {
|
switch v.Kind() {
|
||||||
case reflect.Map:
|
case reflect.Map:
|
||||||
mk := reflect.ValueOf(string(key.Node().Data))
|
vt := v.Type()
|
||||||
|
|
||||||
keyType := v.Type().Key()
|
mk := reflect.ValueOf(string(key.Node().Data))
|
||||||
if !mk.Type().AssignableTo(keyType) {
|
mkt := stringType
|
||||||
if !mk.Type().ConvertibleTo(keyType) {
|
|
||||||
return reflect.Value{}, fmt.Errorf("toml: cannot convert map key of type %s to expected type %s", mk.Type(), keyType)
|
keyType := vt.Key()
|
||||||
|
if !mkt.AssignableTo(keyType) {
|
||||||
|
if !mkt.ConvertibleTo(keyType) {
|
||||||
|
return reflect.Value{}, fmt.Errorf("toml: cannot convert map key of type %s to expected type %s", mkt, keyType)
|
||||||
}
|
}
|
||||||
|
|
||||||
mk = mk.Convert(keyType)
|
mk = mk.Convert(keyType)
|
||||||
@@ -922,7 +1028,7 @@ func (d *decoder) handleKeyValuePart(key ast.Iterator, value *ast.Node, v reflec
|
|||||||
|
|
||||||
// If the map does not exist, create it.
|
// If the map does not exist, create it.
|
||||||
if v.IsNil() {
|
if v.IsNil() {
|
||||||
v = reflect.MakeMap(v.Type())
|
v = reflect.MakeMap(vt)
|
||||||
rv = v
|
rv = v
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -952,12 +1058,20 @@ func (d *decoder) handleKeyValuePart(key ast.Iterator, value *ast.Node, v reflec
|
|||||||
v.SetMapIndex(mk, mv)
|
v.SetMapIndex(mk, mv)
|
||||||
}
|
}
|
||||||
case reflect.Struct:
|
case reflect.Struct:
|
||||||
f, found := structField(v, string(key.Node().Data))
|
path, found := structFieldPath(v, string(key.Node().Data))
|
||||||
if !found {
|
if !found {
|
||||||
d.skipUntilTable = true
|
d.skipUntilTable = true
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if d.errorContext == nil {
|
||||||
|
d.errorContext = new(errorContext)
|
||||||
|
}
|
||||||
|
t := v.Type()
|
||||||
|
d.errorContext.Struct = t
|
||||||
|
d.errorContext.Field = path
|
||||||
|
|
||||||
|
f := fieldByIndex(v, path)
|
||||||
x, err := d.handleKeyValueInner(key, value, f)
|
x, err := d.handleKeyValueInner(key, value, f)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return reflect.Value{}, err
|
return reflect.Value{}, err
|
||||||
@@ -966,14 +1080,17 @@ func (d *decoder) handleKeyValuePart(key ast.Iterator, value *ast.Node, v reflec
|
|||||||
if x.IsValid() {
|
if x.IsValid() {
|
||||||
f.Set(x)
|
f.Set(x)
|
||||||
}
|
}
|
||||||
|
d.errorContext.Struct = nil
|
||||||
|
d.errorContext.Field = nil
|
||||||
case reflect.Interface:
|
case reflect.Interface:
|
||||||
v = v.Elem()
|
v = v.Elem()
|
||||||
|
|
||||||
// Following encoding/toml: decoding an object into an interface{}, it
|
// Following encoding/json: decoding an object into an
|
||||||
// needs to always hold a map[string]interface{}. This is for the types
|
// interface{}, it needs to always hold a
|
||||||
// to be consistent whether a previous value was set or not.
|
// map[string]interface{}. This is for the types to be
|
||||||
|
// consistent whether a previous value was set or not.
|
||||||
if !v.IsValid() || v.Type() != mapStringInterfaceType {
|
if !v.IsValid() || v.Type() != mapStringInterfaceType {
|
||||||
v = reflect.MakeMap(mapStringInterfaceType)
|
v = makeMapStringInterface()
|
||||||
}
|
}
|
||||||
|
|
||||||
x, err := d.handleKeyValuePart(key, value, v)
|
x, err := d.handleKeyValuePart(key, value, v)
|
||||||
@@ -1018,82 +1135,93 @@ func initAndDereferencePointer(v reflect.Value) reflect.Value {
|
|||||||
return elem
|
return elem
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Same as reflect.Value.FieldByIndex, but creates pointers if needed.
|
||||||
|
func fieldByIndex(v reflect.Value, path []int) reflect.Value {
|
||||||
|
for i, x := range path {
|
||||||
|
v = v.Field(x)
|
||||||
|
|
||||||
|
if i < len(path)-1 && v.Kind() == reflect.Pointer {
|
||||||
|
if v.IsNil() {
|
||||||
|
v.Set(reflect.New(v.Type().Elem()))
|
||||||
|
}
|
||||||
|
v = v.Elem()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return v
|
||||||
|
}
|
||||||
|
|
||||||
type fieldPathsMap = map[string][]int
|
type fieldPathsMap = map[string][]int
|
||||||
|
|
||||||
type fieldPathsCache struct {
|
var globalFieldPathsCache atomic.Value // map[danger.TypeID]fieldPathsMap
|
||||||
m map[reflect.Type]fieldPathsMap
|
|
||||||
l sync.RWMutex
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c *fieldPathsCache) get(t reflect.Type) (fieldPathsMap, bool) {
|
func structFieldPath(v reflect.Value, name string) ([]int, bool) {
|
||||||
c.l.RLock()
|
t := v.Type()
|
||||||
paths, ok := c.m[t]
|
|
||||||
c.l.RUnlock()
|
|
||||||
|
|
||||||
return paths, ok
|
cache, _ := globalFieldPathsCache.Load().(map[danger.TypeID]fieldPathsMap)
|
||||||
}
|
fieldPaths, ok := cache[danger.MakeTypeID(t)]
|
||||||
|
|
||||||
func (c *fieldPathsCache) set(t reflect.Type, m fieldPathsMap) {
|
|
||||||
c.l.Lock()
|
|
||||||
c.m[t] = m
|
|
||||||
c.l.Unlock()
|
|
||||||
}
|
|
||||||
|
|
||||||
var globalFieldPathsCache = fieldPathsCache{
|
|
||||||
m: map[reflect.Type]fieldPathsMap{},
|
|
||||||
l: sync.RWMutex{},
|
|
||||||
}
|
|
||||||
|
|
||||||
func structField(v reflect.Value, name string) (reflect.Value, bool) {
|
|
||||||
//nolint:godox
|
|
||||||
// TODO: cache this, and reduce allocations
|
|
||||||
fieldPaths, ok := globalFieldPathsCache.get(v.Type())
|
|
||||||
if !ok {
|
if !ok {
|
||||||
fieldPaths = map[string][]int{}
|
fieldPaths = map[string][]int{}
|
||||||
|
|
||||||
path := make([]int, 0, 16)
|
forEachField(t, nil, func(name string, path []int) {
|
||||||
|
fieldPaths[name] = path
|
||||||
|
// extra copy for the case-insensitive match
|
||||||
|
fieldPaths[strings.ToLower(name)] = path
|
||||||
|
})
|
||||||
|
|
||||||
var walk func(reflect.Value)
|
newCache := make(map[danger.TypeID]fieldPathsMap, len(cache)+1)
|
||||||
walk = func(v reflect.Value) {
|
newCache[danger.MakeTypeID(t)] = fieldPaths
|
||||||
t := v.Type()
|
for k, v := range cache {
|
||||||
for i := 0; i < t.NumField(); i++ {
|
newCache[k] = v
|
||||||
l := len(path)
|
|
||||||
path = append(path, i)
|
|
||||||
f := t.Field(i)
|
|
||||||
|
|
||||||
if f.Anonymous {
|
|
||||||
walk(v.Field(i))
|
|
||||||
} else if f.PkgPath == "" {
|
|
||||||
// only consider exported fields
|
|
||||||
fieldName, ok := f.Tag.Lookup("toml")
|
|
||||||
if !ok {
|
|
||||||
fieldName = f.Name
|
|
||||||
}
|
|
||||||
|
|
||||||
pathCopy := make([]int, len(path))
|
|
||||||
copy(pathCopy, path)
|
|
||||||
|
|
||||||
fieldPaths[fieldName] = pathCopy
|
|
||||||
// extra copy for the case-insensitive match
|
|
||||||
fieldPaths[strings.ToLower(fieldName)] = pathCopy
|
|
||||||
}
|
|
||||||
path = path[:l]
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
globalFieldPathsCache.Store(newCache)
|
||||||
walk(v)
|
|
||||||
|
|
||||||
globalFieldPathsCache.set(v.Type(), fieldPaths)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
path, ok := fieldPaths[name]
|
path, ok := fieldPaths[name]
|
||||||
if !ok {
|
if !ok {
|
||||||
path, ok = fieldPaths[strings.ToLower(name)]
|
path, ok = fieldPaths[strings.ToLower(name)]
|
||||||
}
|
}
|
||||||
|
return path, ok
|
||||||
if !ok {
|
}
|
||||||
return reflect.Value{}, false
|
|
||||||
}
|
func forEachField(t reflect.Type, path []int, do func(name string, path []int)) {
|
||||||
|
n := t.NumField()
|
||||||
return v.FieldByIndex(path), true
|
for i := 0; i < n; i++ {
|
||||||
|
f := t.Field(i)
|
||||||
|
|
||||||
|
if !f.Anonymous && f.PkgPath != "" {
|
||||||
|
// only consider exported fields.
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
fieldPath := append(path, i)
|
||||||
|
fieldPath = fieldPath[:len(fieldPath):len(fieldPath)]
|
||||||
|
|
||||||
|
name := f.Tag.Get("toml")
|
||||||
|
if name == "-" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
if i := strings.IndexByte(name, ','); i >= 0 {
|
||||||
|
name = name[:i]
|
||||||
|
}
|
||||||
|
|
||||||
|
if f.Anonymous && name == "" {
|
||||||
|
t2 := f.Type
|
||||||
|
if t2.Kind() == reflect.Pointer {
|
||||||
|
t2 = t2.Elem()
|
||||||
|
}
|
||||||
|
|
||||||
|
if t2.Kind() == reflect.Struct {
|
||||||
|
forEachField(t2, fieldPath, do)
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
if name == "" {
|
||||||
|
name = f.Name
|
||||||
|
}
|
||||||
|
|
||||||
|
do(name, fieldPath)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
+1354
-232
File diff suppressed because it is too large
Load Diff
@@ -0,0 +1,240 @@
|
|||||||
|
package toml
|
||||||
|
|
||||||
|
import (
|
||||||
|
"unicode/utf8"
|
||||||
|
)
|
||||||
|
|
||||||
|
type utf8Err struct {
|
||||||
|
Index int
|
||||||
|
Size int
|
||||||
|
}
|
||||||
|
|
||||||
|
func (u utf8Err) Zero() bool {
|
||||||
|
return u.Size == 0
|
||||||
|
}
|
||||||
|
|
||||||
|
// Verified that a given string is only made of valid UTF-8 characters allowed
|
||||||
|
// by the TOML spec:
|
||||||
|
//
|
||||||
|
// Any Unicode character may be used except those that must be escaped:
|
||||||
|
// quotation mark, backslash, and the control characters other than tab (U+0000
|
||||||
|
// to U+0008, U+000A to U+001F, U+007F).
|
||||||
|
//
|
||||||
|
// It is a copy of the Go 1.17 utf8.Valid implementation, tweaked to exit early
|
||||||
|
// when a character is not allowed.
|
||||||
|
//
|
||||||
|
// The returned utf8Err is Zero() if the string is valid, or contains the byte
|
||||||
|
// index and size of the invalid character.
|
||||||
|
//
|
||||||
|
// quotation mark => already checked
|
||||||
|
// backslash => already checked
|
||||||
|
// 0-0x8 => invalid
|
||||||
|
// 0x9 => tab, ok
|
||||||
|
// 0xA - 0x1F => invalid
|
||||||
|
// 0x7F => invalid
|
||||||
|
func utf8TomlValidAlreadyEscaped(p []byte) (err utf8Err) {
|
||||||
|
// Fast path. Check for and skip 8 bytes of ASCII characters per iteration.
|
||||||
|
offset := 0
|
||||||
|
for len(p) >= 8 {
|
||||||
|
// Combining two 32 bit loads allows the same code to be used
|
||||||
|
// for 32 and 64 bit platforms.
|
||||||
|
// The compiler can generate a 32bit load for first32 and second32
|
||||||
|
// on many platforms. See test/codegen/memcombine.go.
|
||||||
|
first32 := uint32(p[0]) | uint32(p[1])<<8 | uint32(p[2])<<16 | uint32(p[3])<<24
|
||||||
|
second32 := uint32(p[4]) | uint32(p[5])<<8 | uint32(p[6])<<16 | uint32(p[7])<<24
|
||||||
|
if (first32|second32)&0x80808080 != 0 {
|
||||||
|
// Found a non ASCII byte (>= RuneSelf).
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
for i, b := range p[:8] {
|
||||||
|
if invalidAscii(b) {
|
||||||
|
err.Index = offset + i
|
||||||
|
err.Size = 1
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
p = p[8:]
|
||||||
|
offset += 8
|
||||||
|
}
|
||||||
|
n := len(p)
|
||||||
|
for i := 0; i < n; {
|
||||||
|
pi := p[i]
|
||||||
|
if pi < utf8.RuneSelf {
|
||||||
|
if invalidAscii(pi) {
|
||||||
|
err.Index = offset + i
|
||||||
|
err.Size = 1
|
||||||
|
return
|
||||||
|
}
|
||||||
|
i++
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
x := first[pi]
|
||||||
|
if x == xx {
|
||||||
|
// Illegal starter byte.
|
||||||
|
err.Index = offset + i
|
||||||
|
err.Size = 1
|
||||||
|
return
|
||||||
|
}
|
||||||
|
size := int(x & 7)
|
||||||
|
if i+size > n {
|
||||||
|
// Short or invalid.
|
||||||
|
err.Index = offset + i
|
||||||
|
err.Size = n - i
|
||||||
|
return
|
||||||
|
}
|
||||||
|
accept := acceptRanges[x>>4]
|
||||||
|
if c := p[i+1]; c < accept.lo || accept.hi < c {
|
||||||
|
err.Index = offset + i
|
||||||
|
err.Size = 2
|
||||||
|
return
|
||||||
|
} else if size == 2 {
|
||||||
|
} else if c := p[i+2]; c < locb || hicb < c {
|
||||||
|
err.Index = offset + i
|
||||||
|
err.Size = 3
|
||||||
|
return
|
||||||
|
} else if size == 3 {
|
||||||
|
} else if c := p[i+3]; c < locb || hicb < c {
|
||||||
|
err.Index = offset + i
|
||||||
|
err.Size = 4
|
||||||
|
return
|
||||||
|
}
|
||||||
|
i += size
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Return the size of the next rune if valid, 0 otherwise.
|
||||||
|
func utf8ValidNext(p []byte) int {
|
||||||
|
c := p[0]
|
||||||
|
|
||||||
|
if c < utf8.RuneSelf {
|
||||||
|
if invalidAscii(c) {
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
return 1
|
||||||
|
}
|
||||||
|
|
||||||
|
x := first[c]
|
||||||
|
if x == xx {
|
||||||
|
// Illegal starter byte.
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
size := int(x & 7)
|
||||||
|
if size > len(p) {
|
||||||
|
// Short or invalid.
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
accept := acceptRanges[x>>4]
|
||||||
|
if c := p[1]; c < accept.lo || accept.hi < c {
|
||||||
|
return 0
|
||||||
|
} else if size == 2 {
|
||||||
|
} else if c := p[2]; c < locb || hicb < c {
|
||||||
|
return 0
|
||||||
|
} else if size == 3 {
|
||||||
|
} else if c := p[3]; c < locb || hicb < c {
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
|
||||||
|
return size
|
||||||
|
}
|
||||||
|
|
||||||
|
var invalidAsciiTable = [256]bool{
|
||||||
|
0x00: true,
|
||||||
|
0x01: true,
|
||||||
|
0x02: true,
|
||||||
|
0x03: true,
|
||||||
|
0x04: true,
|
||||||
|
0x05: true,
|
||||||
|
0x06: true,
|
||||||
|
0x07: true,
|
||||||
|
0x08: true,
|
||||||
|
// 0x09 TAB
|
||||||
|
// 0x0A LF
|
||||||
|
0x0B: true,
|
||||||
|
0x0C: true,
|
||||||
|
// 0x0D CR
|
||||||
|
0x0E: true,
|
||||||
|
0x0F: true,
|
||||||
|
0x10: true,
|
||||||
|
0x11: true,
|
||||||
|
0x12: true,
|
||||||
|
0x13: true,
|
||||||
|
0x14: true,
|
||||||
|
0x15: true,
|
||||||
|
0x16: true,
|
||||||
|
0x17: true,
|
||||||
|
0x18: true,
|
||||||
|
0x19: true,
|
||||||
|
0x1A: true,
|
||||||
|
0x1B: true,
|
||||||
|
0x1C: true,
|
||||||
|
0x1D: true,
|
||||||
|
0x1E: true,
|
||||||
|
0x1F: true,
|
||||||
|
// 0x20 - 0x7E Printable ASCII characters
|
||||||
|
0x7F: true,
|
||||||
|
}
|
||||||
|
|
||||||
|
func invalidAscii(b byte) bool {
|
||||||
|
return invalidAsciiTable[b]
|
||||||
|
}
|
||||||
|
|
||||||
|
// acceptRange gives the range of valid values for the second byte in a UTF-8
|
||||||
|
// sequence.
|
||||||
|
type acceptRange struct {
|
||||||
|
lo uint8 // lowest value for second byte.
|
||||||
|
hi uint8 // highest value for second byte.
|
||||||
|
}
|
||||||
|
|
||||||
|
// acceptRanges has size 16 to avoid bounds checks in the code that uses it.
|
||||||
|
var acceptRanges = [16]acceptRange{
|
||||||
|
0: {locb, hicb},
|
||||||
|
1: {0xA0, hicb},
|
||||||
|
2: {locb, 0x9F},
|
||||||
|
3: {0x90, hicb},
|
||||||
|
4: {locb, 0x8F},
|
||||||
|
}
|
||||||
|
|
||||||
|
// first is information about the first byte in a UTF-8 sequence.
|
||||||
|
var first = [256]uint8{
|
||||||
|
// 1 2 3 4 5 6 7 8 9 A B C D E F
|
||||||
|
as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, // 0x00-0x0F
|
||||||
|
as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, // 0x10-0x1F
|
||||||
|
as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, // 0x20-0x2F
|
||||||
|
as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, // 0x30-0x3F
|
||||||
|
as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, // 0x40-0x4F
|
||||||
|
as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, // 0x50-0x5F
|
||||||
|
as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, // 0x60-0x6F
|
||||||
|
as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, as, // 0x70-0x7F
|
||||||
|
// 1 2 3 4 5 6 7 8 9 A B C D E F
|
||||||
|
xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, // 0x80-0x8F
|
||||||
|
xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, // 0x90-0x9F
|
||||||
|
xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, // 0xA0-0xAF
|
||||||
|
xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, // 0xB0-0xBF
|
||||||
|
xx, xx, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, // 0xC0-0xCF
|
||||||
|
s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, s1, // 0xD0-0xDF
|
||||||
|
s2, s3, s3, s3, s3, s3, s3, s3, s3, s3, s3, s3, s3, s4, s3, s3, // 0xE0-0xEF
|
||||||
|
s5, s6, s6, s6, s7, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, xx, // 0xF0-0xFF
|
||||||
|
}
|
||||||
|
|
||||||
|
const (
|
||||||
|
// The default lowest and highest continuation byte.
|
||||||
|
locb = 0b10000000
|
||||||
|
hicb = 0b10111111
|
||||||
|
|
||||||
|
// These names of these constants are chosen to give nice alignment in the
|
||||||
|
// table below. The first nibble is an index into acceptRanges or F for
|
||||||
|
// special one-byte cases. The second nibble is the Rune length or the
|
||||||
|
// Status for the special one-byte case.
|
||||||
|
xx = 0xF1 // invalid: size 1
|
||||||
|
as = 0xF0 // ASCII: size 1
|
||||||
|
s1 = 0x02 // accept 0, size 2
|
||||||
|
s2 = 0x13 // accept 1, size 3
|
||||||
|
s3 = 0x03 // accept 0, size 3
|
||||||
|
s4 = 0x23 // accept 2, size 3
|
||||||
|
s5 = 0x34 // accept 3, size 4
|
||||||
|
s6 = 0x04 // accept 0, size 4
|
||||||
|
s7 = 0x44 // accept 4, size 4
|
||||||
|
)
|
||||||
Reference in New Issue
Block a user