Compare commits
27 Commits
less-reflect
...
raw
| Author | SHA1 | Date | |
|---|---|---|---|
| b371733c67 | |||
| 64dcce07ea | |||
| 28f1efc7d3 | |||
| 7d69e4a728 | |||
| e46d245c09 | |||
| 7baa23f493 | |||
| 2d8433b69e | |||
| 67bc5422f3 | |||
| fb6d1d6c2b | |||
| d017a6dc89 | |||
| d6d3196163 | |||
| 41718a6db3 | |||
| 216628222f | |||
| 322e0b15d2 | |||
| 85bfc0ed51 | |||
| 295a720dfb | |||
| 0a422e3dbd | |||
| 627dade0c7 | |||
| b2e0231cc9 | |||
| ba95863cd3 | |||
| db679df765 | |||
| c5ca2c682b | |||
| ed80712cb4 | |||
| b24772942d | |||
| 9501a05ed7 | |||
| 171a592663 | |||
| 5aaf5ef13b |
@@ -35,11 +35,11 @@ jobs:
|
||||
|
||||
steps:
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v2
|
||||
uses: actions/checkout@v3
|
||||
|
||||
# Initializes the CodeQL tools for scanning.
|
||||
- name: Initialize CodeQL
|
||||
uses: github/codeql-action/init@v1
|
||||
uses: github/codeql-action/init@v2
|
||||
with:
|
||||
languages: ${{ matrix.language }}
|
||||
# If you wish to specify custom queries, you can do so here or in a config file.
|
||||
@@ -50,7 +50,7 @@ jobs:
|
||||
# Autobuild attempts to build any compiled languages (C/C++, C#, or Java).
|
||||
# If this step fails, then you should remove it and run the build manually (see below)
|
||||
- name: Autobuild
|
||||
uses: github/codeql-action/autobuild@v1
|
||||
uses: github/codeql-action/autobuild@v2
|
||||
|
||||
# ℹ️ Command-line programs to run using the OS shell.
|
||||
# 📚 https://git.io/JvXDl
|
||||
@@ -64,4 +64,4 @@ jobs:
|
||||
# make release
|
||||
|
||||
- name: Perform CodeQL Analysis
|
||||
uses: github/codeql-action/analyze@v1
|
||||
uses: github/codeql-action/analyze@v2
|
||||
|
||||
@@ -9,12 +9,12 @@ jobs:
|
||||
runs-on: "ubuntu-latest"
|
||||
name: report
|
||||
steps:
|
||||
- uses: actions/checkout@master
|
||||
- uses: actions/checkout@v3
|
||||
with:
|
||||
fetch-depth: 0
|
||||
- name: Setup go
|
||||
uses: actions/setup-go@master
|
||||
with:
|
||||
go-version: 1.18
|
||||
go-version: 1.19
|
||||
- name: Run tests with coverage
|
||||
run: ./ci.sh coverage -d "${GITHUB_BASE_REF-HEAD}"
|
||||
|
||||
@@ -16,21 +16,21 @@ jobs:
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Checkout
|
||||
uses: actions/checkout@v2
|
||||
uses: actions/checkout@v3
|
||||
with:
|
||||
fetch-depth: 0
|
||||
- name: Set up Go
|
||||
uses: actions/setup-go@v2
|
||||
with:
|
||||
go-version: 1.18
|
||||
go-version: 1.19
|
||||
- name: Login to GitHub Container Registry
|
||||
uses: docker/login-action@v1
|
||||
uses: docker/login-action@v2
|
||||
with:
|
||||
registry: ghcr.io
|
||||
username: ${{ github.actor }}
|
||||
password: ${{ secrets.GITHUB_TOKEN }}
|
||||
- name: Run GoReleaser
|
||||
uses: goreleaser/goreleaser-action@v2
|
||||
uses: goreleaser/goreleaser-action@v3
|
||||
with:
|
||||
distribution: goreleaser
|
||||
version: latest
|
||||
|
||||
@@ -12,11 +12,11 @@ jobs:
|
||||
strategy:
|
||||
matrix:
|
||||
os: [ 'ubuntu-latest', 'windows-latest', 'macos-latest']
|
||||
go: [ '1.17', '1.18' ]
|
||||
go: [ '1.18', '1.19' ]
|
||||
runs-on: ${{ matrix.os }}
|
||||
name: ${{ matrix.go }}/${{ matrix.os }}
|
||||
steps:
|
||||
- uses: actions/checkout@master
|
||||
- uses: actions/checkout@v3
|
||||
with:
|
||||
fetch-depth: 0
|
||||
- name: Setup go ${{ matrix.go }}
|
||||
|
||||
@@ -16,7 +16,11 @@ builds:
|
||||
mod_timestamp: '{{ .CommitTimestamp }}'
|
||||
targets:
|
||||
- linux_amd64
|
||||
- linux_arm64
|
||||
- linux_arm
|
||||
- windows_amd64
|
||||
- windows_arm64
|
||||
- windows_arm
|
||||
- darwin_amd64
|
||||
- darwin_arm64
|
||||
- id: tomljson
|
||||
@@ -31,7 +35,11 @@ builds:
|
||||
mod_timestamp: '{{ .CommitTimestamp }}'
|
||||
targets:
|
||||
- linux_amd64
|
||||
- linux_arm64
|
||||
- linux_arm
|
||||
- windows_amd64
|
||||
- windows_arm64
|
||||
- windows_arm
|
||||
- darwin_amd64
|
||||
- darwin_arm64
|
||||
- id: jsontoml
|
||||
@@ -46,7 +54,11 @@ builds:
|
||||
mod_timestamp: '{{ .CommitTimestamp }}'
|
||||
targets:
|
||||
- linux_amd64
|
||||
- linux_arm64
|
||||
- linux_arm
|
||||
- windows_amd64
|
||||
- windows_arm64
|
||||
- windows_arm
|
||||
- darwin_amd64
|
||||
- darwin_arm64
|
||||
universal_binaries:
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
The MIT License (MIT)
|
||||
|
||||
Copyright (c) 2013 - 2021 Thomas Pelletier, Eric Anderton
|
||||
Copyright (c) 2013 - 2022 Thomas Pelletier, Eric Anderton
|
||||
|
||||
Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||
of this software and associated documentation files (the "Software"), to deal
|
||||
|
||||
@@ -4,17 +4,6 @@ Go library for the [TOML](https://toml.io/en/) format.
|
||||
|
||||
This library supports [TOML v1.0.0](https://toml.io/en/v1.0.0).
|
||||
|
||||
## Development status
|
||||
|
||||
This is the upcoming major version of go-toml. It is currently in active
|
||||
development. As of release v2.0.0-beta.1, the library has reached feature parity
|
||||
with v1, and fixes a lot known bugs and performance issues along the way.
|
||||
|
||||
If you do not need the advanced document editing features of v1, you are
|
||||
encouraged to try out this version.
|
||||
|
||||
[👉 Roadmap for v2](https://github.com/pelletier/go-toml/discussions/506)
|
||||
|
||||
[🐞 Bug Reports](https://github.com/pelletier/go-toml/issues)
|
||||
|
||||
[💬 Anything else](https://github.com/pelletier/go-toml/discussions)
|
||||
@@ -49,7 +38,7 @@ operations should not be shockingly slow. See [benchmarks](#benchmarks).
|
||||
### Strict mode
|
||||
|
||||
`Decoder` can be set to "strict mode", which makes it error when some parts of
|
||||
the TOML document was not prevent in the target structure. This is a great way
|
||||
the TOML document was not present in the target structure. This is a great way
|
||||
to check for typos. [See example in the documentation][strict].
|
||||
|
||||
[strict]: https://pkg.go.dev/github.com/pelletier/go-toml/v2#example-Decoder.DisallowUnknownFields
|
||||
@@ -551,6 +540,13 @@ complete solutions exist out there.
|
||||
[query]: https://github.com/pelletier/go-toml/tree/f99d6bbca119636aeafcf351ee52b3d202782627/query
|
||||
[dasel]: https://github.com/TomWright/dasel
|
||||
|
||||
## Versioning
|
||||
|
||||
Go-toml follows [Semantic Versioning](http://semver.org/). The supported version
|
||||
of [TOML](https://github.com/toml-lang/toml) is indicated at the beginning of
|
||||
this document. The last two major versions of Go are supported
|
||||
(see [Go Release Policy](https://golang.org/doc/devel/release.html#policy)).
|
||||
|
||||
## License
|
||||
|
||||
The MIT License (MIT). Read [LICENSE](LICENSE).
|
||||
|
||||
@@ -1,20 +1,20 @@
|
||||
// Package jsontoml is a program that converts JSON to TOML.
|
||||
//
|
||||
// Usage
|
||||
// # Usage
|
||||
//
|
||||
// Reading from stdin:
|
||||
//
|
||||
// cat file.json | jsontoml > file.toml
|
||||
// cat file.json | jsontoml > file.toml
|
||||
//
|
||||
// Reading from a file:
|
||||
//
|
||||
// jsontoml file.json > file.toml
|
||||
// jsontoml file.json > file.toml
|
||||
//
|
||||
// Installation
|
||||
// # Installation
|
||||
//
|
||||
// Using Go:
|
||||
//
|
||||
// go install github.com/pelletier/go-toml/v2/cmd/jsontoml@latest
|
||||
// go install github.com/pelletier/go-toml/v2/cmd/jsontoml@latest
|
||||
package main
|
||||
|
||||
import (
|
||||
|
||||
@@ -26,7 +26,6 @@ func TestConvert(t *testing.T) {
|
||||
}`,
|
||||
expected: `[mytoml]
|
||||
a = 42.0
|
||||
|
||||
`,
|
||||
},
|
||||
{
|
||||
|
||||
@@ -1,20 +1,20 @@
|
||||
// Package tomljson is a program that converts TOML to JSON.
|
||||
//
|
||||
// Usage
|
||||
// # Usage
|
||||
//
|
||||
// Reading from stdin:
|
||||
//
|
||||
// cat file.toml | tomljson > file.json
|
||||
// cat file.toml | tomljson > file.json
|
||||
//
|
||||
// Reading from a file:
|
||||
//
|
||||
// tomljson file.toml > file.json
|
||||
// tomljson file.toml > file.json
|
||||
//
|
||||
// Installation
|
||||
// # Installation
|
||||
//
|
||||
// Using Go:
|
||||
//
|
||||
// go install github.com/pelletier/go-toml/v2/cmd/tomljson@latest
|
||||
// go install github.com/pelletier/go-toml/v2/cmd/tomljson@latest
|
||||
package main
|
||||
|
||||
import (
|
||||
|
||||
+5
-5
@@ -1,20 +1,20 @@
|
||||
// Package tomll is a linter program for TOML.
|
||||
//
|
||||
// Usage
|
||||
// # Usage
|
||||
//
|
||||
// Reading from stdin, writing to stdout:
|
||||
//
|
||||
// cat file.toml | tomll
|
||||
// cat file.toml | tomll
|
||||
//
|
||||
// Reading and updating a list of files in place:
|
||||
//
|
||||
// tomll a.toml b.toml c.toml
|
||||
// tomll a.toml b.toml c.toml
|
||||
//
|
||||
// Installation
|
||||
// # Installation
|
||||
//
|
||||
// Using Go:
|
||||
//
|
||||
// go install github.com/pelletier/go-toml/v2/cmd/tomll@latest
|
||||
// go install github.com/pelletier/go-toml/v2/cmd/tomll@latest
|
||||
package main
|
||||
|
||||
import (
|
||||
|
||||
@@ -23,7 +23,6 @@ mytoml.a = 42.0
|
||||
`,
|
||||
expected: `[mytoml]
|
||||
a = 42.0
|
||||
|
||||
`,
|
||||
},
|
||||
{
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
//
|
||||
// Within the go-toml package, run `go generate`. Otherwise, use:
|
||||
//
|
||||
// go run github.com/pelletier/go-toml/cmd/tomltestgen -o toml_testgen_test.go
|
||||
// go run github.com/pelletier/go-toml/cmd/tomltestgen -o toml_testgen_test.go
|
||||
package main
|
||||
|
||||
import (
|
||||
|
||||
@@ -103,6 +103,7 @@ func (e *DecodeError) Key() Key {
|
||||
//
|
||||
// The function copies all bytes used in DecodeError, so that document and
|
||||
// highlight can be freely deallocated.
|
||||
//
|
||||
//nolint:funlen
|
||||
func wrapDecodeError(document []byte, de *decodeError) *DecodeError {
|
||||
offset := danger.SubsliceOffset(document, de.highlight)
|
||||
|
||||
+2
-2
@@ -1,5 +1,5 @@
|
||||
//go:build go1.18
|
||||
// +build go1.18
|
||||
//go:build go1.18 || go1.19
|
||||
// +build go1.18 go1.19
|
||||
|
||||
package toml_test
|
||||
|
||||
|
||||
@@ -2,4 +2,4 @@ module github.com/pelletier/go-toml/v2
|
||||
|
||||
go 1.16
|
||||
|
||||
require github.com/stretchr/testify v1.7.1
|
||||
require github.com/stretchr/testify v1.8.0
|
||||
|
||||
@@ -1,11 +1,15 @@
|
||||
github.com/davecgh/go-spew v1.1.0 h1:ZDRjVQ15GmhC3fiQ8ni8+OwkZQO4DARzQgrnXU1Liz8=
|
||||
github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
||||
github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c=
|
||||
github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
||||
github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM=
|
||||
github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4=
|
||||
github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
||||
github.com/stretchr/testify v1.7.1 h1:5TQK59W5E3v0r2duFAb7P95B6hEeOyEnHRa8MjYSMTY=
|
||||
github.com/stretchr/objx v0.4.0/go.mod h1:YvHI0jy2hoMjB+UWwv71VJQ9isScKT/TqJzVSSt89Yw=
|
||||
github.com/stretchr/testify v1.7.1/go.mod h1:6Fq8oRcR53rry900zMqJjRRixrwX3KX962/h/Wwjteg=
|
||||
github.com/stretchr/testify v1.8.0 h1:pSgiaMZlXftHpm5L7V1+rVB+AZJydKsMxsQBIJw4PKk=
|
||||
github.com/stretchr/testify v1.8.0/go.mod h1:yNjHg4UonilssWZ8iaSj1OCr/vHnekPRkoO+kdMU+MU=
|
||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405 h1:yhCVgyC4o1eVCa2tZl7eS0r+SDo693bJlVdllGtEeKM=
|
||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
||||
gopkg.in/yaml.v3 v3.0.0-20200313102051-9f266ea9e77c h1:dUUwHk2QECo/6vqA44rthZ8ie2QXMNeKRTHCNY2nXvo=
|
||||
gopkg.in/yaml.v3 v3.0.0-20200313102051-9f266ea9e77c/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
|
||||
gopkg.in/yaml.v3 v3.0.1 h1:fxVm/GzAzEWqLHuvctI91KS9hhNmmWOoWu0XTYJS7CA=
|
||||
gopkg.in/yaml.v3 v3.0.1/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
|
||||
|
||||
+4
-4
@@ -11,10 +11,10 @@ import (
|
||||
//
|
||||
// For example:
|
||||
//
|
||||
// it := n.Children()
|
||||
// for it.Next() {
|
||||
// it.Node()
|
||||
// }
|
||||
// it := n.Children()
|
||||
// for it.Next() {
|
||||
// it.Node()
|
||||
// }
|
||||
type Iterator struct {
|
||||
started bool
|
||||
node *Node
|
||||
|
||||
@@ -67,6 +67,7 @@ func TestDocMarshal(t *testing.T) {
|
||||
}
|
||||
|
||||
marshalTestToml := `title = 'TOML Marshal Testing'
|
||||
|
||||
[basic_lists]
|
||||
floats = [12.3, 45.6, 78.9]
|
||||
bools = [true, false, true]
|
||||
@@ -89,7 +90,6 @@ name = 'Second'
|
||||
[subdoc.first]
|
||||
name = 'First'
|
||||
|
||||
|
||||
[basic]
|
||||
uint = 5001
|
||||
bool = true
|
||||
@@ -101,9 +101,9 @@ date = 1979-05-27T07:32:00Z
|
||||
|
||||
[[subdoclist]]
|
||||
name = 'List.First'
|
||||
|
||||
[[subdoclist]]
|
||||
name = 'List.Second'
|
||||
|
||||
`
|
||||
|
||||
result, err := toml.Marshal(docData)
|
||||
@@ -117,14 +117,15 @@ func TestBasicMarshalQuotedKey(t *testing.T) {
|
||||
|
||||
expected := `'Z.string-àéù' = 'Hello'
|
||||
'Yfloat-𝟘' = 3.5
|
||||
|
||||
['Xsubdoc-àéù']
|
||||
String2 = 'One'
|
||||
|
||||
[['W.sublist-𝟘']]
|
||||
String2 = 'Two'
|
||||
|
||||
[['W.sublist-𝟘']]
|
||||
String2 = 'Three'
|
||||
|
||||
`
|
||||
|
||||
require.Equal(t, string(expected), string(result))
|
||||
@@ -159,8 +160,8 @@ bool = false
|
||||
int = 0
|
||||
string = ''
|
||||
stringlist = []
|
||||
[map]
|
||||
|
||||
[map]
|
||||
`
|
||||
|
||||
require.Equal(t, string(expected), string(result))
|
||||
|
||||
@@ -151,6 +151,7 @@ type quotedKeyMarshalTestStruct struct {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var quotedKeyMarshalTestData = quotedKeyMarshalTestStruct{
|
||||
String: "Hello",
|
||||
@@ -160,6 +161,7 @@ var quotedKeyMarshalTestData = quotedKeyMarshalTestStruct{
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var quotedKeyMarshalTestToml = []byte(`"Yfloat-𝟘" = 3.5
|
||||
"Z.string-àéù" = "Hello"
|
||||
@@ -272,6 +274,7 @@ var docData = testDoc{
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var mapTestDoc = testMapDoc{
|
||||
Title: "TOML Marshal Testing",
|
||||
@@ -559,10 +562,12 @@ func (c customMarshaler) MarshalTOML() ([]byte, error) {
|
||||
var customMarshalerData = customMarshaler{FirstName: "Sally", LastName: "Fields"}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var customMarshalerToml = []byte(`Sally Fields`)
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var nestedCustomMarshalerData = customMarshalerParent{
|
||||
Self: customMarshaler{FirstName: "Maiku", LastName: "Suteda"},
|
||||
@@ -570,6 +575,7 @@ var nestedCustomMarshalerData = customMarshalerParent{
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var nestedCustomMarshalerToml = []byte(`friends = ["Sally Fields"]
|
||||
me = "Maiku Suteda"
|
||||
@@ -611,6 +617,7 @@ func TestUnmarshalTextMarshaler(t *testing.T) {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once type and methods are used by a test
|
||||
//
|
||||
//nolint:unused
|
||||
type precedentMarshaler struct {
|
||||
FirstName string
|
||||
@@ -629,6 +636,7 @@ func (m precedentMarshaler) MarshalTOML() ([]byte, error) {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once type and method are used by a test
|
||||
//
|
||||
//nolint:unused
|
||||
type customPointerMarshaler struct {
|
||||
FirstName string
|
||||
@@ -641,6 +649,7 @@ func (m *customPointerMarshaler) MarshalTOML() ([]byte, error) {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once type and method are used by a test
|
||||
//
|
||||
//nolint:unused
|
||||
type textPointerMarshaler struct {
|
||||
FirstName string
|
||||
@@ -653,6 +662,7 @@ func (m *textPointerMarshaler) MarshalText() ([]byte, error) {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var commentTestToml = []byte(`
|
||||
# it's a comment on type
|
||||
@@ -690,6 +700,7 @@ type mapsTestStruct struct {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var mapsTestData = mapsTestStruct{
|
||||
Simple: map[string]string{
|
||||
@@ -713,6 +724,7 @@ var mapsTestData = mapsTestStruct{
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var mapsTestToml = []byte(`
|
||||
[Other]
|
||||
@@ -735,6 +747,7 @@ var mapsTestToml = []byte(`
|
||||
`)
|
||||
|
||||
// TODO: Remove nolint once type is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused
|
||||
type structArrayNoTag struct {
|
||||
A struct {
|
||||
@@ -744,6 +757,7 @@ type structArrayNoTag struct {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var customTagTestToml = []byte(`
|
||||
[postgres]
|
||||
@@ -758,6 +772,7 @@ var customTagTestToml = []byte(`
|
||||
`)
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var customCommentTagTestToml = []byte(`
|
||||
# db connection
|
||||
@@ -771,6 +786,7 @@ var customCommentTagTestToml = []byte(`
|
||||
`)
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var customCommentedTagTestToml = []byte(`
|
||||
[postgres]
|
||||
@@ -825,6 +841,7 @@ func TestUnmarshalTabInStringAndQuotedKey(t *testing.T) {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var customMultilineTagTestToml = []byte(`int_slice = [
|
||||
1,
|
||||
@@ -834,6 +851,7 @@ var customMultilineTagTestToml = []byte(`int_slice = [
|
||||
`)
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var testDocBasicToml = []byte(`
|
||||
[document]
|
||||
@@ -846,12 +864,14 @@ var testDocBasicToml = []byte(`
|
||||
`)
|
||||
|
||||
// TODO: Remove nolint once type is used by a test
|
||||
//
|
||||
//nolint:deadcode
|
||||
type testDocCustomTag struct {
|
||||
Doc testDocBasicsCustomTag `file:"document"`
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once type is used by a test
|
||||
//
|
||||
//nolint:deadcode
|
||||
type testDocBasicsCustomTag struct {
|
||||
Bool bool `file:"bool_val"`
|
||||
@@ -864,6 +884,7 @@ type testDocBasicsCustomTag struct {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,varcheck
|
||||
var testDocCustomTagData = testDocCustomTag{
|
||||
Doc: testDocBasicsCustomTag{
|
||||
@@ -966,6 +987,7 @@ func TestUnmarshalInvalidPointerKind(t *testing.T) {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused
|
||||
type testDuration struct {
|
||||
Nanosec time.Duration `toml:"nanosec"`
|
||||
@@ -980,6 +1002,7 @@ type testDuration struct {
|
||||
}
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var testDurationToml = []byte(`
|
||||
nanosec = "1ns"
|
||||
@@ -994,6 +1017,7 @@ a_string = "15s"
|
||||
`)
|
||||
|
||||
// TODO: Remove nolint once var is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused,varcheck
|
||||
var testDurationToml2 = []byte(`a_string = "15s"
|
||||
hour = "1h0m0s"
|
||||
@@ -1007,6 +1031,7 @@ sec = "1s"
|
||||
`)
|
||||
|
||||
// TODO: Remove nolint once type is used by a test
|
||||
//
|
||||
//nolint:deadcode,unused
|
||||
type testBadDuration struct {
|
||||
Val time.Duration `toml:"val"`
|
||||
|
||||
@@ -11,5 +11,6 @@ func TestEntrySize(t *testing.T) {
|
||||
// Validate no regression on the size of entry{}. This is a critical bit for
|
||||
// performance of unmarshaling documents. Should only be increased with care
|
||||
// and a very good reason.
|
||||
require.LessOrEqual(t, 48, int(unsafe.Sizeof(entry{})))
|
||||
maxExpectedEntrySize := 48
|
||||
require.LessOrEqual(t, int(unsafe.Sizeof(entry{})), maxExpectedEntrySize)
|
||||
}
|
||||
|
||||
+106
-16
@@ -54,7 +54,7 @@ func NewEncoder(w io.Writer) *Encoder {
|
||||
// This behavior can be controlled on an individual struct field basis with the
|
||||
// inline tag:
|
||||
//
|
||||
// MyField `inline:"true"`
|
||||
// MyField `toml:",inline"`
|
||||
func (enc *Encoder) SetTablesInline(inline bool) *Encoder {
|
||||
enc.tablesInline = inline
|
||||
return enc
|
||||
@@ -65,7 +65,7 @@ func (enc *Encoder) SetTablesInline(inline bool) *Encoder {
|
||||
//
|
||||
// This behavior can be controlled on an individual struct field basis with the multiline tag:
|
||||
//
|
||||
// MyField `multiline:"true"`
|
||||
// MyField `multiline:"true"`
|
||||
func (enc *Encoder) SetArraysMultiline(multiline bool) *Encoder {
|
||||
enc.arraysMultiline = multiline
|
||||
return enc
|
||||
@@ -89,7 +89,7 @@ func (enc *Encoder) SetIndentTables(indent bool) *Encoder {
|
||||
//
|
||||
// If v cannot be represented to TOML it returns an error.
|
||||
//
|
||||
// Encoding rules
|
||||
// # Encoding rules
|
||||
//
|
||||
// A top level slice containing only maps or structs is encoded as [[table
|
||||
// array]].
|
||||
@@ -107,10 +107,30 @@ func (enc *Encoder) SetIndentTables(indent bool) *Encoder {
|
||||
// a newline character or a single quote. In that case they are emitted as
|
||||
// quoted strings.
|
||||
//
|
||||
// Unsigned integers larger than math.MaxInt64 cannot be encoded. Doing so
|
||||
// results in an error. This rule exists because the TOML specification only
|
||||
// requires parsers to support at least the 64 bits integer range. Allowing
|
||||
// larger numbers would create non-standard TOML documents, which may not be
|
||||
// readable (at best) by other implementations. To encode such numbers, a
|
||||
// solution is a custom type that implements encoding.TextMarshaler.
|
||||
//
|
||||
// When encoding structs, fields are encoded in order of definition, with their
|
||||
// exact name.
|
||||
//
|
||||
// Struct tags
|
||||
// Tables and array tables are separated by empty lines. However, consecutive
|
||||
// subtables definitions are not. For example:
|
||||
//
|
||||
// [top1]
|
||||
//
|
||||
// [top2]
|
||||
// [top2.child1]
|
||||
//
|
||||
// [[array]]
|
||||
//
|
||||
// [[array]]
|
||||
// [array.child2]
|
||||
//
|
||||
// # Struct tags
|
||||
//
|
||||
// The encoding of each public struct field can be customized by the format
|
||||
// string in the "toml" key of the struct field's tag. This follows
|
||||
@@ -128,7 +148,8 @@ func (enc *Encoder) SetIndentTables(indent bool) *Encoder {
|
||||
//
|
||||
// In addition to the "toml" tag struct tag, a "comment" tag can be used to emit
|
||||
// a TOML comment before the value being annotated. Comments are ignored inside
|
||||
// inline tables.
|
||||
// inline tables. For array tables, the comment is only present before the first
|
||||
// element of the array.
|
||||
func (enc *Encoder) Encode(v interface{}) error {
|
||||
var (
|
||||
b []byte
|
||||
@@ -302,7 +323,11 @@ func (enc *Encoder) encode(b []byte, ctx encoderCtx, v reflect.Value) ([]byte, e
|
||||
b = append(b, "false"...)
|
||||
}
|
||||
case reflect.Uint64, reflect.Uint32, reflect.Uint16, reflect.Uint8, reflect.Uint:
|
||||
b = strconv.AppendUint(b, v.Uint(), 10)
|
||||
x := v.Uint()
|
||||
if x > uint64(math.MaxInt64) {
|
||||
return nil, fmt.Errorf("toml: not encoding uint (%d) greater than max int64 (%d)", x, int64(math.MaxInt64))
|
||||
}
|
||||
b = strconv.AppendUint(b, x, 10)
|
||||
case reflect.Int64, reflect.Int32, reflect.Int16, reflect.Int8, reflect.Int:
|
||||
b = strconv.AppendInt(b, v.Int(), 10)
|
||||
default:
|
||||
@@ -321,13 +346,13 @@ func isNil(v reflect.Value) bool {
|
||||
}
|
||||
}
|
||||
|
||||
func shouldOmitEmpty(options valueOptions, v reflect.Value) bool {
|
||||
return options.omitempty && isEmptyValue(v)
|
||||
}
|
||||
|
||||
func (enc *Encoder) encodeKv(b []byte, ctx encoderCtx, options valueOptions, v reflect.Value) ([]byte, error) {
|
||||
var err error
|
||||
|
||||
if (ctx.options.omitempty || options.omitempty) && isEmptyValue(v) {
|
||||
return b, nil
|
||||
}
|
||||
|
||||
if !ctx.inline {
|
||||
b = enc.encodeComment(ctx.indent, options.comment, b)
|
||||
}
|
||||
@@ -353,6 +378,8 @@ func (enc *Encoder) encodeKv(b []byte, ctx encoderCtx, options valueOptions, v r
|
||||
|
||||
func isEmptyValue(v reflect.Value) bool {
|
||||
switch v.Kind() {
|
||||
case reflect.Struct:
|
||||
return isEmptyStruct(v)
|
||||
case reflect.Array, reflect.Map, reflect.Slice, reflect.String:
|
||||
return v.Len() == 0
|
||||
case reflect.Bool:
|
||||
@@ -369,6 +396,34 @@ func isEmptyValue(v reflect.Value) bool {
|
||||
return false
|
||||
}
|
||||
|
||||
func isEmptyStruct(v reflect.Value) bool {
|
||||
// TODO: merge with walkStruct and cache.
|
||||
typ := v.Type()
|
||||
for i := 0; i < typ.NumField(); i++ {
|
||||
fieldType := typ.Field(i)
|
||||
|
||||
// only consider exported fields
|
||||
if fieldType.PkgPath != "" {
|
||||
continue
|
||||
}
|
||||
|
||||
tag := fieldType.Tag.Get("toml")
|
||||
|
||||
// special field name to skip field
|
||||
if tag == "-" {
|
||||
continue
|
||||
}
|
||||
|
||||
f := v.Field(i)
|
||||
|
||||
if !isEmptyValue(f) {
|
||||
return false
|
||||
}
|
||||
}
|
||||
|
||||
return true
|
||||
}
|
||||
|
||||
const literalQuote = '\''
|
||||
|
||||
func (enc *Encoder) encodeString(b []byte, v string, options valueOptions) []byte {
|
||||
@@ -398,7 +453,6 @@ func (enc *Encoder) encodeLiteralString(b []byte, v string) []byte {
|
||||
return b
|
||||
}
|
||||
|
||||
//nolint:cyclop
|
||||
func (enc *Encoder) encodeQuotedString(multiline bool, b []byte, v string) []byte {
|
||||
stringQuote := `"`
|
||||
|
||||
@@ -652,10 +706,19 @@ func (enc *Encoder) encodeStruct(b []byte, ctx encoderCtx, v reflect.Value) ([]b
|
||||
}
|
||||
|
||||
func (enc *Encoder) encodeComment(indent int, comment string, b []byte) []byte {
|
||||
if comment != "" {
|
||||
for len(comment) > 0 {
|
||||
var line string
|
||||
idx := strings.IndexByte(comment, '\n')
|
||||
if idx >= 0 {
|
||||
line = comment[:idx]
|
||||
comment = comment[idx+1:]
|
||||
} else {
|
||||
line = comment
|
||||
comment = ""
|
||||
}
|
||||
b = enc.indent(indent, b)
|
||||
b = append(b, "# "...)
|
||||
b = append(b, comment...)
|
||||
b = append(b, line...)
|
||||
b = append(b, '\n')
|
||||
}
|
||||
return b
|
||||
@@ -736,7 +799,13 @@ func (enc *Encoder) encodeTable(b []byte, ctx encoderCtx, t table) ([]byte, erro
|
||||
}
|
||||
ctx.skipTableHeader = false
|
||||
|
||||
hasNonEmptyKV := false
|
||||
for _, kv := range t.kvs {
|
||||
if shouldOmitEmpty(kv.Options, kv.Value) {
|
||||
continue
|
||||
}
|
||||
hasNonEmptyKV = true
|
||||
|
||||
ctx.setKey(kv.Key)
|
||||
|
||||
b, err = enc.encodeKv(b, ctx, kv.Options, kv.Value)
|
||||
@@ -747,7 +816,20 @@ func (enc *Encoder) encodeTable(b []byte, ctx encoderCtx, t table) ([]byte, erro
|
||||
b = append(b, '\n')
|
||||
}
|
||||
|
||||
first := true
|
||||
for _, table := range t.tables {
|
||||
if shouldOmitEmpty(table.Options, table.Value) {
|
||||
continue
|
||||
}
|
||||
if first {
|
||||
first = false
|
||||
if hasNonEmptyKV {
|
||||
b = append(b, '\n')
|
||||
}
|
||||
} else {
|
||||
b = append(b, "\n"...)
|
||||
}
|
||||
|
||||
ctx.setKey(table.Key)
|
||||
|
||||
ctx.options = table.Options
|
||||
@@ -756,8 +838,6 @@ func (enc *Encoder) encodeTable(b []byte, ctx encoderCtx, t table) ([]byte, erro
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
b = append(b, '\n')
|
||||
}
|
||||
|
||||
return b, nil
|
||||
@@ -770,6 +850,10 @@ func (enc *Encoder) encodeTableInline(b []byte, ctx encoderCtx, t table) ([]byte
|
||||
|
||||
first := true
|
||||
for _, kv := range t.kvs {
|
||||
if shouldOmitEmpty(kv.Options, kv.Value) {
|
||||
continue
|
||||
}
|
||||
|
||||
if first {
|
||||
first = false
|
||||
} else {
|
||||
@@ -785,7 +869,7 @@ func (enc *Encoder) encodeTableInline(b []byte, ctx encoderCtx, t table) ([]byte
|
||||
}
|
||||
|
||||
if len(t.tables) > 0 {
|
||||
panic("inline table cannot contain nested tables, online key-values")
|
||||
panic("inline table cannot contain nested tables, only key-values")
|
||||
}
|
||||
|
||||
b = append(b, "}"...)
|
||||
@@ -881,7 +965,13 @@ func (enc *Encoder) encodeSliceAsArrayTable(b []byte, ctx encoderCtx, v reflect.
|
||||
scratch = append(scratch, "]]\n"...)
|
||||
ctx.skipTableHeader = true
|
||||
|
||||
b = enc.encodeComment(ctx.indent, ctx.options.comment, b)
|
||||
|
||||
for i := 0; i < v.Len(); i++ {
|
||||
if i != 0 {
|
||||
b = append(b, "\n"...)
|
||||
}
|
||||
|
||||
b = append(b, scratch...)
|
||||
|
||||
var err error
|
||||
|
||||
+197
-93
@@ -39,21 +39,21 @@ func TestMarshal(t *testing.T) {
|
||||
v: map[string]string{
|
||||
"hello": "world",
|
||||
},
|
||||
expected: "hello = 'world'",
|
||||
expected: "hello = 'world'\n",
|
||||
},
|
||||
{
|
||||
desc: "map with new line in key",
|
||||
v: map[string]string{
|
||||
"hel\nlo": "world",
|
||||
},
|
||||
expected: `"hel\nlo" = 'world'`,
|
||||
expected: "\"hel\\nlo\" = 'world'\n",
|
||||
},
|
||||
{
|
||||
desc: `map with " in key`,
|
||||
v: map[string]string{
|
||||
`hel"lo`: "world",
|
||||
},
|
||||
expected: `'hel"lo' = 'world'`,
|
||||
expected: "'hel\"lo' = 'world'\n",
|
||||
},
|
||||
{
|
||||
desc: "map in map and string",
|
||||
@@ -62,9 +62,9 @@ func TestMarshal(t *testing.T) {
|
||||
"hello": "world",
|
||||
},
|
||||
},
|
||||
expected: `
|
||||
[table]
|
||||
hello = 'world'`,
|
||||
expected: `[table]
|
||||
hello = 'world'
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "map in map in map and string",
|
||||
@@ -75,10 +75,10 @@ hello = 'world'`,
|
||||
},
|
||||
},
|
||||
},
|
||||
expected: `
|
||||
[this]
|
||||
expected: `[this]
|
||||
[this.is]
|
||||
a = 'test'`,
|
||||
a = 'test'
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "map in map in map and string with values",
|
||||
@@ -90,18 +90,20 @@ a = 'test'`,
|
||||
"also": "that",
|
||||
},
|
||||
},
|
||||
expected: `
|
||||
[this]
|
||||
expected: `[this]
|
||||
also = 'that'
|
||||
|
||||
[this.is]
|
||||
a = 'test'`,
|
||||
a = 'test'
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "simple string array",
|
||||
v: map[string][]string{
|
||||
"array": {"one", "two", "three"},
|
||||
},
|
||||
expected: `array = ['one', 'two', 'three']`,
|
||||
expected: `array = ['one', 'two', 'three']
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "empty string array",
|
||||
@@ -118,14 +120,16 @@ a = 'test'`,
|
||||
v: map[string][][]string{
|
||||
"array": {{"one", "two"}, {"three"}},
|
||||
},
|
||||
expected: `array = [['one', 'two'], ['three']]`,
|
||||
expected: `array = [['one', 'two'], ['three']]
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "mixed strings and nested string arrays",
|
||||
v: map[string][]interface{}{
|
||||
"array": {"a string", []string{"one", "two"}, "last"},
|
||||
},
|
||||
expected: `array = ['a string', ['one', 'two'], 'last']`,
|
||||
expected: `array = ['a string', ['one', 'two'], 'last']
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "array of maps",
|
||||
@@ -135,9 +139,9 @@ a = 'test'`,
|
||||
{"map2.1": "v2.1"},
|
||||
},
|
||||
},
|
||||
expected: `
|
||||
[[top]]
|
||||
expected: `[[top]]
|
||||
'map1.1' = 'v1.1'
|
||||
|
||||
[[top]]
|
||||
'map2.1' = 'v2.1'
|
||||
`,
|
||||
@@ -148,9 +152,9 @@ a = 'test'`,
|
||||
"key1": "value1",
|
||||
"key2": "value2",
|
||||
},
|
||||
expected: `
|
||||
key1 = 'value1'
|
||||
key2 = 'value2'`,
|
||||
expected: `key1 = 'value1'
|
||||
key2 = 'value2'
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "simple struct",
|
||||
@@ -159,7 +163,8 @@ key2 = 'value2'`,
|
||||
}{
|
||||
A: "foo",
|
||||
},
|
||||
expected: `A = 'foo'`,
|
||||
expected: `A = 'foo'
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "one level of structs within structs",
|
||||
@@ -174,8 +179,7 @@ key2 = 'value2'`,
|
||||
K2: "v2",
|
||||
},
|
||||
},
|
||||
expected: `
|
||||
[A]
|
||||
expected: `[A]
|
||||
K1 = 'v1'
|
||||
K2 = 'v2'
|
||||
`,
|
||||
@@ -190,10 +194,10 @@ K2 = 'v2'
|
||||
},
|
||||
},
|
||||
},
|
||||
expected: `
|
||||
[root]
|
||||
expected: `[root]
|
||||
[[root.nested]]
|
||||
name = 'Bob'
|
||||
|
||||
[[root.nested]]
|
||||
name = 'Alice'
|
||||
`,
|
||||
@@ -203,49 +207,53 @@ name = 'Alice'
|
||||
v: map[string]interface{}{
|
||||
"a": "'\b\f\r\t\"\\",
|
||||
},
|
||||
expected: `a = "'\b\f\r\t\"\\"`,
|
||||
expected: `a = "'\b\f\r\t\"\\"
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "string utf8 low",
|
||||
v: map[string]interface{}{
|
||||
"a": "'Ę",
|
||||
},
|
||||
expected: `a = "'Ę"`,
|
||||
expected: `a = "'Ę"
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "string utf8 low 2",
|
||||
v: map[string]interface{}{
|
||||
"a": "'\u10A85",
|
||||
},
|
||||
expected: "a = \"'\u10A85\"",
|
||||
expected: "a = \"'\u10A85\"\n",
|
||||
},
|
||||
{
|
||||
desc: "string utf8 low 2",
|
||||
v: map[string]interface{}{
|
||||
"a": "'\u10A85",
|
||||
},
|
||||
expected: "a = \"'\u10A85\"",
|
||||
expected: "a = \"'\u10A85\"\n",
|
||||
},
|
||||
{
|
||||
desc: "emoji",
|
||||
v: map[string]interface{}{
|
||||
"a": "'😀",
|
||||
},
|
||||
expected: "a = \"'😀\"",
|
||||
expected: "a = \"'😀\"\n",
|
||||
},
|
||||
{
|
||||
desc: "control char",
|
||||
v: map[string]interface{}{
|
||||
"a": "'\u001A",
|
||||
},
|
||||
expected: `a = "'\u001A"`,
|
||||
expected: `a = "'\u001A"
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "multi-line string",
|
||||
v: map[string]interface{}{
|
||||
"a": "hello\nworld",
|
||||
},
|
||||
expected: `a = "hello\nworld"`,
|
||||
expected: `a = "hello\nworld"
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "multi-line forced",
|
||||
@@ -256,7 +264,8 @@ name = 'Alice'
|
||||
},
|
||||
expected: `A = """
|
||||
hello
|
||||
world"""`,
|
||||
world"""
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "inline field",
|
||||
@@ -271,8 +280,8 @@ world"""`,
|
||||
"isinline": "no",
|
||||
},
|
||||
},
|
||||
expected: `
|
||||
A = {isinline = 'yes'}
|
||||
expected: `A = {isinline = 'yes'}
|
||||
|
||||
[B]
|
||||
isinline = 'no'
|
||||
`,
|
||||
@@ -286,8 +295,7 @@ isinline = 'no'
|
||||
A: []int{1, 2, 3, 4},
|
||||
B: []int{1, 2, 3, 4},
|
||||
},
|
||||
expected: `
|
||||
A = [
|
||||
expected: `A = [
|
||||
1,
|
||||
2,
|
||||
3,
|
||||
@@ -303,8 +311,7 @@ B = [1, 2, 3, 4]
|
||||
}{
|
||||
A: [][]int{{1, 2}, {3, 4}},
|
||||
},
|
||||
expected: `
|
||||
A = [
|
||||
expected: `A = [
|
||||
[1, 2],
|
||||
[3, 4]
|
||||
]
|
||||
@@ -329,7 +336,8 @@ A = [
|
||||
}{
|
||||
A: []*int{nil},
|
||||
},
|
||||
expected: `A = [0]`,
|
||||
expected: `A = [0]
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "nil pointer in slice uses zero value",
|
||||
@@ -338,7 +346,8 @@ A = [
|
||||
}{
|
||||
A: []*int{nil},
|
||||
},
|
||||
expected: `A = [0]`,
|
||||
expected: `A = [0]
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "pointer in slice",
|
||||
@@ -347,7 +356,8 @@ A = [
|
||||
}{
|
||||
A: []*int{&someInt},
|
||||
},
|
||||
expected: `A = [42]`,
|
||||
expected: `A = [42]
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "inline table in inline table",
|
||||
@@ -358,23 +368,25 @@ A = [
|
||||
},
|
||||
},
|
||||
},
|
||||
expected: `A = {A = {A = 'hello'}}`,
|
||||
expected: `A = {A = {A = 'hello'}}
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "empty slice in map",
|
||||
v: map[string][]string{
|
||||
"a": {},
|
||||
},
|
||||
expected: `a = []`,
|
||||
expected: `a = []
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "map in slice",
|
||||
v: map[string][]map[string]string{
|
||||
"a": {{"hello": "world"}},
|
||||
},
|
||||
expected: `
|
||||
[[a]]
|
||||
hello = 'world'`,
|
||||
expected: `[[a]]
|
||||
hello = 'world'
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "newline in map in slice",
|
||||
@@ -382,7 +394,8 @@ hello = 'world'`,
|
||||
"a\n": {{"hello": "world"}},
|
||||
},
|
||||
expected: `[["a\n"]]
|
||||
hello = 'world'`,
|
||||
hello = 'world'
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "newline in map in slice",
|
||||
@@ -398,7 +411,8 @@ hello = 'world'`,
|
||||
}{
|
||||
A: []struct{}{},
|
||||
},
|
||||
expected: `A = []`,
|
||||
expected: `A = []
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "nil field is ignored",
|
||||
@@ -418,7 +432,8 @@ hello = 'world'`,
|
||||
Public: "shown",
|
||||
private: "hidden",
|
||||
},
|
||||
expected: `Public = 'shown'`,
|
||||
expected: `Public = 'shown'
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "fields tagged - are ignored",
|
||||
@@ -442,7 +457,8 @@ hello = 'world'`,
|
||||
v: map[string]interface{}{
|
||||
"hello\nworld": 42,
|
||||
},
|
||||
expected: `"hello\nworld" = 42`,
|
||||
expected: `"hello\nworld" = 42
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "new line in parent of nested table key",
|
||||
@@ -452,7 +468,8 @@ hello = 'world'`,
|
||||
},
|
||||
},
|
||||
expected: `["hello\nworld"]
|
||||
inner = 42`,
|
||||
inner = 42
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "new line in nested table key",
|
||||
@@ -465,7 +482,8 @@ inner = 42`,
|
||||
},
|
||||
expected: `[parent]
|
||||
[parent."in\ner"]
|
||||
foo = 42`,
|
||||
foo = 42
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "invalid map key",
|
||||
@@ -488,7 +506,8 @@ foo = 42`,
|
||||
}{
|
||||
T: time.Time{},
|
||||
},
|
||||
expected: `T = 0001-01-01T00:00:00Z`,
|
||||
expected: `T = 0001-01-01T00:00:00Z
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "time nano",
|
||||
@@ -497,7 +516,8 @@ foo = 42`,
|
||||
}{
|
||||
T: time.Date(1979, time.May, 27, 0, 32, 0, 999999000, time.UTC),
|
||||
},
|
||||
expected: `T = 1979-05-27T00:32:00.999999Z`,
|
||||
expected: `T = 1979-05-27T00:32:00.999999Z
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "bool",
|
||||
@@ -508,9 +528,9 @@ foo = 42`,
|
||||
A: false,
|
||||
B: true,
|
||||
},
|
||||
expected: `
|
||||
A = false
|
||||
B = true`,
|
||||
expected: `A = false
|
||||
B = true
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "numbers",
|
||||
@@ -541,8 +561,7 @@ B = true`,
|
||||
K: 42,
|
||||
L: 2.2,
|
||||
},
|
||||
expected: `
|
||||
A = 1.1
|
||||
expected: `A = 1.1
|
||||
B = 42
|
||||
C = 42
|
||||
D = 42
|
||||
@@ -553,7 +572,8 @@ H = 42
|
||||
I = 42
|
||||
J = 42
|
||||
K = 42
|
||||
L = 2.2`,
|
||||
L = 2.2
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "comments",
|
||||
@@ -566,8 +586,7 @@ L = 2.2`,
|
||||
Three: []int{1, 2, 3},
|
||||
},
|
||||
},
|
||||
expected: `
|
||||
# Before table
|
||||
expected: `# Before table
|
||||
[Table]
|
||||
One = 1
|
||||
# Before kv
|
||||
@@ -589,7 +608,7 @@ Three = [1, 2, 3]
|
||||
}
|
||||
|
||||
require.NoError(t, err)
|
||||
equalStringsIgnoreNewlines(t, e.expected, string(b))
|
||||
assert.Equal(t, e.expected, string(b))
|
||||
|
||||
// make sure the output is always valid TOML
|
||||
defaultMap := map[string]interface{}{}
|
||||
@@ -664,12 +683,6 @@ func testWithFlags(t *testing.T, flags int, setters flagsSetters, testfn func(t
|
||||
}
|
||||
}
|
||||
|
||||
func equalStringsIgnoreNewlines(t *testing.T, expected string, actual string) {
|
||||
t.Helper()
|
||||
cutset := "\n"
|
||||
assert.Equal(t, strings.Trim(expected, cutset), strings.Trim(actual, cutset))
|
||||
}
|
||||
|
||||
func TestMarshalFloats(t *testing.T) {
|
||||
v := map[string]float32{
|
||||
"nan": float32(math.NaN()),
|
||||
@@ -709,7 +722,8 @@ func TestMarshalIndentTables(t *testing.T) {
|
||||
v: map[string]interface{}{
|
||||
"foo": "bar",
|
||||
},
|
||||
expected: `foo = 'bar'`,
|
||||
expected: `foo = 'bar'
|
||||
`,
|
||||
},
|
||||
{
|
||||
desc: "one level table",
|
||||
@@ -719,8 +733,7 @@ func TestMarshalIndentTables(t *testing.T) {
|
||||
"two": "value2",
|
||||
},
|
||||
},
|
||||
expected: `
|
||||
[foo]
|
||||
expected: `[foo]
|
||||
one = 'value1'
|
||||
two = 'value2'
|
||||
`,
|
||||
@@ -736,10 +749,11 @@ func TestMarshalIndentTables(t *testing.T) {
|
||||
},
|
||||
},
|
||||
},
|
||||
expected: `
|
||||
root = 'value0'
|
||||
expected: `root = 'value0'
|
||||
|
||||
[level1]
|
||||
one = 'value1'
|
||||
|
||||
[level1.level2]
|
||||
two = 'value2'
|
||||
`,
|
||||
@@ -754,7 +768,7 @@ root = 'value0'
|
||||
enc.SetIndentTables(true)
|
||||
err := enc.Encode(e.v)
|
||||
require.NoError(t, err)
|
||||
equalStringsIgnoreNewlines(t, e.expected, buf.String())
|
||||
assert.Equal(t, e.expected, buf.String())
|
||||
})
|
||||
}
|
||||
}
|
||||
@@ -799,7 +813,7 @@ func TestMarshalTextMarshaler(t *testing.T) {
|
||||
m := map[string]interface{}{"a": &customTextMarshaler{value: 2}}
|
||||
r, err := toml.Marshal(m)
|
||||
require.NoError(t, err)
|
||||
equalStringsIgnoreNewlines(t, "a = '::2'", string(r))
|
||||
assert.Equal(t, "a = '::2'\n", string(r))
|
||||
}
|
||||
|
||||
type brokenWriter struct{}
|
||||
@@ -822,10 +836,10 @@ func TestEncoderSetIndentSymbol(t *testing.T) {
|
||||
enc.SetIndentSymbol(">>>")
|
||||
err := enc.Encode(map[string]map[string]string{"parent": {"hello": "world"}})
|
||||
require.NoError(t, err)
|
||||
expected := `
|
||||
[parent]
|
||||
>>>hello = 'world'`
|
||||
equalStringsIgnoreNewlines(t, expected, w.String())
|
||||
expected := `[parent]
|
||||
>>>hello = 'world'
|
||||
`
|
||||
assert.Equal(t, expected, w.String())
|
||||
}
|
||||
|
||||
func TestEncoderOmitempty(t *testing.T) {
|
||||
@@ -856,9 +870,9 @@ func TestEncoderOmitempty(t *testing.T) {
|
||||
b, err := toml.Marshal(d)
|
||||
require.NoError(t, err)
|
||||
|
||||
expected := `[Struct]`
|
||||
expected := ``
|
||||
|
||||
equalStringsIgnoreNewlines(t, expected, string(b))
|
||||
assert.Equal(t, expected, string(b))
|
||||
}
|
||||
|
||||
func TestEncoderTagFieldName(t *testing.T) {
|
||||
@@ -873,13 +887,12 @@ func TestEncoderTagFieldName(t *testing.T) {
|
||||
b, err := toml.Marshal(d)
|
||||
require.NoError(t, err)
|
||||
|
||||
expected := `
|
||||
hello = 'world'
|
||||
expected := `hello = 'world'
|
||||
'#' = ''
|
||||
Bad = ''
|
||||
`
|
||||
|
||||
equalStringsIgnoreNewlines(t, expected, string(b))
|
||||
assert.Equal(t, expected, string(b))
|
||||
}
|
||||
|
||||
func TestIssue436(t *testing.T) {
|
||||
@@ -893,12 +906,11 @@ func TestIssue436(t *testing.T) {
|
||||
err = toml.NewEncoder(&buf).Encode(v)
|
||||
require.NoError(t, err)
|
||||
|
||||
expected := `
|
||||
[[a]]
|
||||
expected := `[[a]]
|
||||
[a.b]
|
||||
c = 'd'
|
||||
`
|
||||
equalStringsIgnoreNewlines(t, expected, buf.String())
|
||||
assert.Equal(t, expected, buf.String())
|
||||
}
|
||||
|
||||
func TestIssue424(t *testing.T) {
|
||||
@@ -980,7 +992,7 @@ func TestIssue678(t *testing.T) {
|
||||
|
||||
out, err := toml.Marshal(cfg)
|
||||
require.NoError(t, err)
|
||||
equalStringsIgnoreNewlines(t, "BigInt = '123'", string(out))
|
||||
assert.Equal(t, "BigInt = '123'\n", string(out))
|
||||
|
||||
cfg2 := &Config{}
|
||||
err = toml.Unmarshal(out, cfg2)
|
||||
@@ -1004,6 +1016,85 @@ func TestIssue752(t *testing.T) {
|
||||
require.Equal(t, "", string(out))
|
||||
}
|
||||
|
||||
func TestIssue768(t *testing.T) {
|
||||
type cfg struct {
|
||||
Name string `comment:"This is a multiline comment.\nThis is line 2."`
|
||||
}
|
||||
|
||||
out, err := toml.Marshal(&cfg{})
|
||||
require.NoError(t, err)
|
||||
|
||||
expected := `# This is a multiline comment.
|
||||
# This is line 2.
|
||||
Name = ''
|
||||
`
|
||||
|
||||
require.Equal(t, expected, string(out))
|
||||
}
|
||||
|
||||
func TestIssue786(t *testing.T) {
|
||||
type Dependencies struct {
|
||||
Dependencies []string `toml:"dependencies,multiline,omitempty"`
|
||||
BuildDependencies []string `toml:"buildDependencies,multiline,omitempty"`
|
||||
OptionalDependencies []string `toml:"optionalDependencies,multiline,omitempty"`
|
||||
}
|
||||
|
||||
type Test struct {
|
||||
Dependencies Dependencies `toml:"dependencies,omitempty"`
|
||||
}
|
||||
|
||||
x := Test{}
|
||||
b, err := toml.Marshal(x)
|
||||
require.NoError(t, err)
|
||||
|
||||
require.Equal(t, "", string(b))
|
||||
|
||||
type General struct {
|
||||
From string `toml:"from,omitempty" json:"from,omitempty" comment:"from in graphite-web format, the local TZ is used"`
|
||||
Randomize bool `toml:"randomize" json:"randomize" comment:"randomize starting time with [0,step)"`
|
||||
}
|
||||
|
||||
type Custom struct {
|
||||
Name string `toml:"name" json:"name,omitempty" comment:"names for generator, braces are expanded like in shell"`
|
||||
Type string `toml:"type,omitempty" json:"type,omitempty" comment:"type of generator"`
|
||||
General
|
||||
}
|
||||
type Config struct {
|
||||
General
|
||||
Custom []Custom `toml:"custom,omitempty" json:"custom,omitempty" comment:"generators with custom parameters can be specified separately"`
|
||||
}
|
||||
|
||||
buf := new(bytes.Buffer)
|
||||
config := &Config{General: General{From: "-2d", Randomize: true}}
|
||||
config.Custom = []Custom{{Name: "omit", General: General{Randomize: false}}}
|
||||
config.Custom = append(config.Custom, Custom{Name: "present", General: General{From: "-2d", Randomize: true}})
|
||||
encoder := toml.NewEncoder(buf)
|
||||
encoder.Encode(config)
|
||||
|
||||
expected := `# from in graphite-web format, the local TZ is used
|
||||
from = '-2d'
|
||||
# randomize starting time with [0,step)
|
||||
randomize = true
|
||||
|
||||
# generators with custom parameters can be specified separately
|
||||
[[custom]]
|
||||
# names for generator, braces are expanded like in shell
|
||||
name = 'omit'
|
||||
# randomize starting time with [0,step)
|
||||
randomize = false
|
||||
|
||||
[[custom]]
|
||||
# names for generator, braces are expanded like in shell
|
||||
name = 'present'
|
||||
# from in graphite-web format, the local TZ is used
|
||||
from = '-2d'
|
||||
# randomize starting time with [0,step)
|
||||
randomize = true
|
||||
`
|
||||
|
||||
require.Equal(t, expected, buf.String())
|
||||
}
|
||||
|
||||
func TestMarshalNestedAnonymousStructs(t *testing.T) {
|
||||
type Embedded struct {
|
||||
Value string `toml:"value" json:"value"`
|
||||
@@ -1025,6 +1116,7 @@ func TestMarshalNestedAnonymousStructs(t *testing.T) {
|
||||
}
|
||||
|
||||
expected := `value = ''
|
||||
|
||||
[top]
|
||||
value = ''
|
||||
|
||||
@@ -1033,7 +1125,6 @@ value = ''
|
||||
|
||||
[anonymous]
|
||||
value = ''
|
||||
|
||||
`
|
||||
|
||||
result, err := toml.Marshal(doc)
|
||||
@@ -1057,9 +1148,9 @@ func TestMarshalNestedAnonymousStructs_DuplicateField(t *testing.T) {
|
||||
doc.Value = "shadows"
|
||||
|
||||
expected := `value = 'shadows'
|
||||
|
||||
[top]
|
||||
value = ''
|
||||
|
||||
`
|
||||
|
||||
result, err := toml.Marshal(doc)
|
||||
@@ -1070,7 +1161,7 @@ value = ''
|
||||
|
||||
func TestLocalTime(t *testing.T) {
|
||||
v := map[string]toml.LocalTime{
|
||||
"a": toml.LocalTime{
|
||||
"a": {
|
||||
Hour: 1,
|
||||
Minute: 2,
|
||||
Second: 3,
|
||||
@@ -1086,6 +1177,19 @@ func TestLocalTime(t *testing.T) {
|
||||
require.Equal(t, expected, string(out))
|
||||
}
|
||||
|
||||
func TestMarshalUint64Overflow(t *testing.T) {
|
||||
// The TOML spec only requires implementation to provide support for the
|
||||
// int64 range. To avoid generating TOML documents that would not be
|
||||
// supported by standard-compliant parsers, uint64 > max int64 cannot be
|
||||
// marshaled.
|
||||
x := map[string]interface{}{
|
||||
"foo": uint64(math.MaxInt64) + 1,
|
||||
}
|
||||
|
||||
_, err := toml.Marshal(x)
|
||||
require.Error(t, err)
|
||||
}
|
||||
|
||||
func ExampleMarshal() {
|
||||
type MyConfig struct {
|
||||
Version int
|
||||
|
||||
@@ -28,6 +28,11 @@ func (p *parser) Raw(raw ast.Range) []byte {
|
||||
return p.data[raw.Offset : raw.Offset+raw.Length]
|
||||
}
|
||||
|
||||
func (p *parser) SetRaw(ref ast.Reference, from []byte, to []byte) {
|
||||
b := danger.BytesRange(from, to)
|
||||
p.builder.NodeAt(ref).Raw = p.Range(b)
|
||||
}
|
||||
|
||||
func (p *parser) Reset(b []byte) {
|
||||
p.builder.Reset()
|
||||
p.ref = ast.InvalidReference
|
||||
@@ -152,12 +157,14 @@ func (p *parser) parseArrayTable(b []byte) (ast.Reference, []byte, error) {
|
||||
ref := p.builder.Push(ast.Node{
|
||||
Kind: ast.ArrayTable,
|
||||
})
|
||||
start := b
|
||||
|
||||
b = b[2:]
|
||||
b = p.parseWhitespace(b)
|
||||
|
||||
k, b, err := p.parseKey(b)
|
||||
if err != nil {
|
||||
p.SetRaw(ref, start, b)
|
||||
return ref, nil, err
|
||||
}
|
||||
|
||||
@@ -166,11 +173,12 @@ func (p *parser) parseArrayTable(b []byte) (ast.Reference, []byte, error) {
|
||||
|
||||
b, err = expect(']', b)
|
||||
if err != nil {
|
||||
p.SetRaw(ref, start, b)
|
||||
return ref, nil, err
|
||||
}
|
||||
|
||||
b, err = expect(']', b)
|
||||
|
||||
p.SetRaw(ref, start, b)
|
||||
return ref, b, err
|
||||
}
|
||||
|
||||
@@ -181,12 +189,14 @@ func (p *parser) parseStdTable(b []byte) (ast.Reference, []byte, error) {
|
||||
ref := p.builder.Push(ast.Node{
|
||||
Kind: ast.Table,
|
||||
})
|
||||
start := b
|
||||
|
||||
b = b[1:]
|
||||
b = p.parseWhitespace(b)
|
||||
|
||||
key, b, err := p.parseKey(b)
|
||||
if err != nil {
|
||||
p.SetRaw(ref, start, b)
|
||||
return ref, nil, err
|
||||
}
|
||||
|
||||
@@ -196,6 +206,7 @@ func (p *parser) parseStdTable(b []byte) (ast.Reference, []byte, error) {
|
||||
|
||||
b, err = expect(']', b)
|
||||
|
||||
p.SetRaw(ref, start, b)
|
||||
return ref, b, err
|
||||
}
|
||||
|
||||
@@ -204,10 +215,12 @@ func (p *parser) parseKeyval(b []byte) (ast.Reference, []byte, error) {
|
||||
ref := p.builder.Push(ast.Node{
|
||||
Kind: ast.KeyValue,
|
||||
})
|
||||
start := b
|
||||
|
||||
key, b, err := p.parseKey(b)
|
||||
if err != nil {
|
||||
return ast.InvalidReference, nil, err
|
||||
p.SetRaw(ref, start, b)
|
||||
return ast.InvalidReference, b, err
|
||||
}
|
||||
|
||||
// keyval-sep = ws %x3D ws ; =
|
||||
@@ -215,24 +228,28 @@ func (p *parser) parseKeyval(b []byte) (ast.Reference, []byte, error) {
|
||||
b = p.parseWhitespace(b)
|
||||
|
||||
if len(b) == 0 {
|
||||
return ast.InvalidReference, nil, newDecodeError(b, "expected = after a key, but the document ends there")
|
||||
p.SetRaw(ref, start, b)
|
||||
return ast.InvalidReference, b, newDecodeError(b, "expected = after a key, but the document ends there")
|
||||
}
|
||||
|
||||
b, err = expect('=', b)
|
||||
if err != nil {
|
||||
return ast.InvalidReference, nil, err
|
||||
p.SetRaw(ref, start, b)
|
||||
return ast.InvalidReference, b, err
|
||||
}
|
||||
|
||||
b = p.parseWhitespace(b)
|
||||
|
||||
valRef, b, err := p.parseVal(b)
|
||||
if err != nil {
|
||||
p.SetRaw(ref, start, b)
|
||||
return ref, b, err
|
||||
}
|
||||
|
||||
p.builder.Chain(valRef, key)
|
||||
p.builder.AttachChild(ref, valRef)
|
||||
|
||||
p.SetRaw(ref, start, b)
|
||||
return ref, b, err
|
||||
}
|
||||
|
||||
@@ -242,7 +259,7 @@ func (p *parser) parseVal(b []byte) (ast.Reference, []byte, error) {
|
||||
ref := ast.InvalidReference
|
||||
|
||||
if len(b) == 0 {
|
||||
return ref, nil, newDecodeError(b, "expected value, not eof")
|
||||
return ref, b, newDecodeError(b, "expected value, not eof")
|
||||
}
|
||||
|
||||
var err error
|
||||
@@ -287,23 +304,25 @@ func (p *parser) parseVal(b []byte) (ast.Reference, []byte, error) {
|
||||
return ref, b, err
|
||||
case 't':
|
||||
if !scanFollowsTrue(b) {
|
||||
return ref, nil, newDecodeError(atmost(b, 4), "expected 'true'")
|
||||
return ref, b, newDecodeError(atmost(b, 4), "expected 'true'")
|
||||
}
|
||||
|
||||
ref = p.builder.Push(ast.Node{
|
||||
Kind: ast.Bool,
|
||||
Data: b[:4],
|
||||
Raw: p.Range(b[:4]),
|
||||
})
|
||||
|
||||
return ref, b[4:], nil
|
||||
case 'f':
|
||||
if !scanFollowsFalse(b) {
|
||||
return ref, nil, newDecodeError(atmost(b, 5), "expected 'false'")
|
||||
return ref, b, newDecodeError(atmost(b, 5), "expected 'false'")
|
||||
}
|
||||
|
||||
ref = p.builder.Push(ast.Node{
|
||||
Kind: ast.Bool,
|
||||
Data: b[:5],
|
||||
Raw: p.Range(b[:5]),
|
||||
})
|
||||
|
||||
return ref, b[5:], nil
|
||||
@@ -327,7 +346,7 @@ func atmost(b []byte, n int) []byte {
|
||||
func (p *parser) parseLiteralString(b []byte) ([]byte, []byte, []byte, error) {
|
||||
v, rest, err := scanLiteralString(b)
|
||||
if err != nil {
|
||||
return nil, nil, nil, err
|
||||
return nil, nil, rest, err
|
||||
}
|
||||
|
||||
return v, v[1 : len(v)-1], rest, nil
|
||||
@@ -342,6 +361,7 @@ func (p *parser) parseInlineTable(b []byte) (ast.Reference, []byte, error) {
|
||||
parent := p.builder.Push(ast.Node{
|
||||
Kind: ast.InlineTable,
|
||||
})
|
||||
start := b
|
||||
|
||||
first := true
|
||||
|
||||
@@ -356,7 +376,8 @@ func (p *parser) parseInlineTable(b []byte) (ast.Reference, []byte, error) {
|
||||
b = p.parseWhitespace(b)
|
||||
|
||||
if len(b) == 0 {
|
||||
return parent, nil, newDecodeError(previousB[:1], "inline table is incomplete")
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, b, newDecodeError(previousB[:1], "inline table is incomplete")
|
||||
}
|
||||
|
||||
if b[0] == '}' {
|
||||
@@ -366,7 +387,8 @@ func (p *parser) parseInlineTable(b []byte) (ast.Reference, []byte, error) {
|
||||
if !first {
|
||||
b, err = expect(',', b)
|
||||
if err != nil {
|
||||
return parent, nil, err
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, b, err
|
||||
}
|
||||
b = p.parseWhitespace(b)
|
||||
}
|
||||
@@ -375,7 +397,8 @@ func (p *parser) parseInlineTable(b []byte) (ast.Reference, []byte, error) {
|
||||
|
||||
kv, b, err = p.parseKeyval(b)
|
||||
if err != nil {
|
||||
return parent, nil, err
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, b, err
|
||||
}
|
||||
|
||||
if first {
|
||||
@@ -390,6 +413,7 @@ func (p *parser) parseInlineTable(b []byte) (ast.Reference, []byte, error) {
|
||||
|
||||
rest, err := expect('}', b)
|
||||
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, rest, err
|
||||
}
|
||||
|
||||
@@ -403,6 +427,7 @@ func (p *parser) parseValArray(b []byte) (ast.Reference, []byte, error) {
|
||||
// array-sep = %x2C ; , Comma
|
||||
// ws-comment-newline = *( wschar / [ comment ] newline )
|
||||
arrayStart := b
|
||||
start := b
|
||||
b = b[1:]
|
||||
|
||||
parent := p.builder.Push(ast.Node{
|
||||
@@ -417,11 +442,13 @@ func (p *parser) parseValArray(b []byte) (ast.Reference, []byte, error) {
|
||||
for len(b) > 0 {
|
||||
b, err = p.parseOptionalWhitespaceCommentNewline(b)
|
||||
if err != nil {
|
||||
return parent, nil, err
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, b, err
|
||||
}
|
||||
|
||||
if len(b) == 0 {
|
||||
return parent, nil, newDecodeError(arrayStart[:1], "array is incomplete")
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, b, newDecodeError(arrayStart[:1], "array is incomplete")
|
||||
}
|
||||
|
||||
if b[0] == ']' {
|
||||
@@ -430,16 +457,19 @@ func (p *parser) parseValArray(b []byte) (ast.Reference, []byte, error) {
|
||||
|
||||
if b[0] == ',' {
|
||||
if first {
|
||||
return parent, nil, newDecodeError(b[0:1], "array cannot start with comma")
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, b, newDecodeError(b[0:1], "array cannot start with comma")
|
||||
}
|
||||
b = b[1:]
|
||||
|
||||
b, err = p.parseOptionalWhitespaceCommentNewline(b)
|
||||
if err != nil {
|
||||
return parent, nil, err
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, b, err
|
||||
}
|
||||
} else if !first {
|
||||
return parent, nil, newDecodeError(b[0:1], "array elements must be separated by commas")
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, b, newDecodeError(b[0:1], "array elements must be separated by commas")
|
||||
}
|
||||
|
||||
// TOML allows trailing commas in arrays.
|
||||
@@ -450,7 +480,8 @@ func (p *parser) parseValArray(b []byte) (ast.Reference, []byte, error) {
|
||||
var valueRef ast.Reference
|
||||
valueRef, b, err = p.parseVal(b)
|
||||
if err != nil {
|
||||
return parent, nil, err
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, b, err
|
||||
}
|
||||
|
||||
if first {
|
||||
@@ -462,13 +493,16 @@ func (p *parser) parseValArray(b []byte) (ast.Reference, []byte, error) {
|
||||
|
||||
b, err = p.parseOptionalWhitespaceCommentNewline(b)
|
||||
if err != nil {
|
||||
return parent, nil, err
|
||||
p.SetRaw(parent, start, b)
|
||||
return parent, b, err
|
||||
}
|
||||
first = false
|
||||
}
|
||||
|
||||
rest, err := expect(']', b)
|
||||
|
||||
p.SetRaw(parent, start, rest)
|
||||
|
||||
return parent, rest, err
|
||||
}
|
||||
|
||||
@@ -480,7 +514,7 @@ func (p *parser) parseOptionalWhitespaceCommentNewline(b []byte) ([]byte, error)
|
||||
if len(b) > 0 && b[0] == '#' {
|
||||
_, b, err = scanComment(b)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
return b, err
|
||||
}
|
||||
}
|
||||
|
||||
@@ -491,7 +525,7 @@ func (p *parser) parseOptionalWhitespaceCommentNewline(b []byte) ([]byte, error)
|
||||
if b[0] == '\n' || b[0] == '\r' {
|
||||
b, err = p.parseNewline(b)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
return b, err
|
||||
}
|
||||
} else {
|
||||
break
|
||||
@@ -504,7 +538,7 @@ func (p *parser) parseOptionalWhitespaceCommentNewline(b []byte) ([]byte, error)
|
||||
func (p *parser) parseMultilineLiteralString(b []byte) ([]byte, []byte, []byte, error) {
|
||||
token, rest, err := scanMultilineLiteralString(b)
|
||||
if err != nil {
|
||||
return nil, nil, nil, err
|
||||
return nil, nil, rest, err
|
||||
}
|
||||
|
||||
i := 3
|
||||
@@ -533,7 +567,7 @@ func (p *parser) parseMultilineBasicString(b []byte) ([]byte, []byte, []byte, er
|
||||
// mlb-escaped-nl = escape ws newline *( wschar / newline )
|
||||
token, escaped, rest, err := scanMultilineBasicString(b)
|
||||
if err != nil {
|
||||
return nil, nil, nil, err
|
||||
return nil, nil, rest, err
|
||||
}
|
||||
|
||||
i := 3
|
||||
@@ -555,7 +589,7 @@ func (p *parser) parseMultilineBasicString(b []byte) ([]byte, []byte, []byte, er
|
||||
if verr.Zero() {
|
||||
return token, str, rest, nil
|
||||
}
|
||||
return nil, nil, nil, newDecodeError(str[verr.Index:verr.Index+verr.Size], "invalid UTF-8")
|
||||
return nil, nil, rest, newDecodeError(str[verr.Index:verr.Index+verr.Size], "invalid UTF-8")
|
||||
}
|
||||
|
||||
var builder bytes.Buffer
|
||||
@@ -622,26 +656,26 @@ func (p *parser) parseMultilineBasicString(b []byte) ([]byte, []byte, []byte, er
|
||||
case 'u':
|
||||
x, err := hexToRune(atmost(token[i+1:], 4), 4)
|
||||
if err != nil {
|
||||
return nil, nil, nil, err
|
||||
return nil, nil, rest, err
|
||||
}
|
||||
builder.WriteRune(x)
|
||||
i += 4
|
||||
case 'U':
|
||||
x, err := hexToRune(atmost(token[i+1:], 8), 8)
|
||||
if err != nil {
|
||||
return nil, nil, nil, err
|
||||
return nil, nil, rest, err
|
||||
}
|
||||
|
||||
builder.WriteRune(x)
|
||||
i += 8
|
||||
default:
|
||||
return nil, nil, nil, newDecodeError(token[i:i+1], "invalid escaped character %#U", c)
|
||||
return nil, nil, rest, newDecodeError(token[i:i+1], "invalid escaped character %#U", c)
|
||||
}
|
||||
i++
|
||||
} else {
|
||||
size := utf8ValidNext(token[i:])
|
||||
if size == 0 {
|
||||
return nil, nil, nil, newDecodeError(token[i:i+1], "invalid character %#U", c)
|
||||
return nil, nil, rest, newDecodeError(token[i:i+1], "invalid character %#U", c)
|
||||
}
|
||||
builder.Write(token[i : i+size])
|
||||
i += size
|
||||
@@ -662,7 +696,7 @@ func (p *parser) parseKey(b []byte) (ast.Reference, []byte, error) {
|
||||
// dot-sep = ws %x2E ws ; . Period
|
||||
raw, key, b, err := p.parseSimpleKey(b)
|
||||
if err != nil {
|
||||
return ast.InvalidReference, nil, err
|
||||
return ast.InvalidReference, b, err
|
||||
}
|
||||
|
||||
ref := p.builder.Push(ast.Node{
|
||||
@@ -678,7 +712,7 @@ func (p *parser) parseKey(b []byte) (ast.Reference, []byte, error) {
|
||||
|
||||
raw, key, b, err = p.parseSimpleKey(b)
|
||||
if err != nil {
|
||||
return ref, nil, err
|
||||
return ref, b, err
|
||||
}
|
||||
|
||||
p.builder.PushAndChain(ast.Node{
|
||||
@@ -696,7 +730,7 @@ func (p *parser) parseKey(b []byte) (ast.Reference, []byte, error) {
|
||||
|
||||
func (p *parser) parseSimpleKey(b []byte) (raw, key, rest []byte, err error) {
|
||||
if len(b) == 0 {
|
||||
return nil, nil, nil, newDecodeError(b, "expected key but found none")
|
||||
return nil, nil, b, newDecodeError(b, "expected key but found none")
|
||||
}
|
||||
|
||||
// simple-key = quoted-key / unquoted-key
|
||||
@@ -711,7 +745,7 @@ func (p *parser) parseSimpleKey(b []byte) (raw, key, rest []byte, err error) {
|
||||
key, rest = scanUnquotedKey(b)
|
||||
return key, key, rest, nil
|
||||
default:
|
||||
return nil, nil, nil, newDecodeError(b[0:1], "invalid character at start of key: %c", b[0])
|
||||
return nil, nil, b[1:], newDecodeError(b[0:1], "invalid character at start of key: %c", b[0])
|
||||
}
|
||||
}
|
||||
|
||||
@@ -733,7 +767,7 @@ func (p *parser) parseBasicString(b []byte) ([]byte, []byte, []byte, error) {
|
||||
// escape-seq-char =/ %x55 8HEXDIG ; UXXXXXXXX U+XXXXXXXX
|
||||
token, escaped, rest, err := scanBasicString(b)
|
||||
if err != nil {
|
||||
return nil, nil, nil, err
|
||||
return nil, nil, rest, err
|
||||
}
|
||||
|
||||
startIdx := len(`"`)
|
||||
@@ -748,7 +782,7 @@ func (p *parser) parseBasicString(b []byte) ([]byte, []byte, []byte, error) {
|
||||
if verr.Zero() {
|
||||
return token, str, rest, nil
|
||||
}
|
||||
return nil, nil, nil, newDecodeError(str[verr.Index:verr.Index+verr.Size], "invalid UTF-8")
|
||||
return nil, nil, rest, newDecodeError(str[verr.Index:verr.Index+verr.Size], "invalid UTF-8")
|
||||
}
|
||||
|
||||
i := startIdx
|
||||
@@ -781,7 +815,7 @@ func (p *parser) parseBasicString(b []byte) ([]byte, []byte, []byte, error) {
|
||||
case 'u':
|
||||
x, err := hexToRune(token[i+1:len(token)-1], 4)
|
||||
if err != nil {
|
||||
return nil, nil, nil, err
|
||||
return nil, nil, rest, err
|
||||
}
|
||||
|
||||
builder.WriteRune(x)
|
||||
@@ -789,19 +823,19 @@ func (p *parser) parseBasicString(b []byte) ([]byte, []byte, []byte, error) {
|
||||
case 'U':
|
||||
x, err := hexToRune(token[i+1:len(token)-1], 8)
|
||||
if err != nil {
|
||||
return nil, nil, nil, err
|
||||
return nil, nil, rest, err
|
||||
}
|
||||
|
||||
builder.WriteRune(x)
|
||||
i += 8
|
||||
default:
|
||||
return nil, nil, nil, newDecodeError(token[i:i+1], "invalid escaped character %#U", c)
|
||||
return nil, nil, rest, newDecodeError(token[i:i+1], "invalid escaped character %#U", c)
|
||||
}
|
||||
i++
|
||||
} else {
|
||||
size := utf8ValidNext(token[i:])
|
||||
if size == 0 {
|
||||
return nil, nil, nil, newDecodeError(token[i:i+1], "invalid character %#U", c)
|
||||
return nil, nil, rest, newDecodeError(token[i:i+1], "invalid character %#U", c)
|
||||
}
|
||||
builder.Write(token[i : i+size])
|
||||
i += size
|
||||
@@ -854,21 +888,23 @@ func (p *parser) parseIntOrFloatOrDateTime(b []byte) (ast.Reference, []byte, err
|
||||
switch b[0] {
|
||||
case 'i':
|
||||
if !scanFollowsInf(b) {
|
||||
return ast.InvalidReference, nil, newDecodeError(atmost(b, 3), "expected 'inf'")
|
||||
return ast.InvalidReference, b, newDecodeError(atmost(b, 3), "expected 'inf'")
|
||||
}
|
||||
|
||||
return p.builder.Push(ast.Node{
|
||||
Kind: ast.Float,
|
||||
Data: b[:3],
|
||||
Raw: p.Range(b[:3]),
|
||||
}), b[3:], nil
|
||||
case 'n':
|
||||
if !scanFollowsNan(b) {
|
||||
return ast.InvalidReference, nil, newDecodeError(atmost(b, 3), "expected 'nan'")
|
||||
return ast.InvalidReference, b, newDecodeError(atmost(b, 3), "expected 'nan'")
|
||||
}
|
||||
|
||||
return p.builder.Push(ast.Node{
|
||||
Kind: ast.Float,
|
||||
Data: b[:3],
|
||||
Raw: p.Range(b[:3]),
|
||||
}), b[3:], nil
|
||||
case '+', '-':
|
||||
return p.scanIntOrFloat(b)
|
||||
@@ -960,6 +996,7 @@ byteLoop:
|
||||
return p.builder.Push(ast.Node{
|
||||
Kind: kind,
|
||||
Data: b[:i],
|
||||
Raw: p.Range(b[:i]),
|
||||
}), b[i:], nil
|
||||
}
|
||||
|
||||
@@ -993,6 +1030,7 @@ func (p *parser) scanIntOrFloat(b []byte) (ast.Reference, []byte, error) {
|
||||
return p.builder.Push(ast.Node{
|
||||
Kind: ast.Integer,
|
||||
Data: b[:i],
|
||||
Raw: p.Range(b[:i]),
|
||||
}), b[i:], nil
|
||||
}
|
||||
|
||||
@@ -1016,10 +1054,11 @@ func (p *parser) scanIntOrFloat(b []byte) (ast.Reference, []byte, error) {
|
||||
return p.builder.Push(ast.Node{
|
||||
Kind: ast.Float,
|
||||
Data: b[:i+3],
|
||||
Raw: p.Range(b[:i+3]),
|
||||
}), b[i+3:], nil
|
||||
}
|
||||
|
||||
return ast.InvalidReference, nil, newDecodeError(b[i:i+1], "unexpected character 'i' while scanning for a number")
|
||||
return ast.InvalidReference, b[i:], newDecodeError(b[i:i+1], "unexpected character 'i' while scanning for a number")
|
||||
}
|
||||
|
||||
if c == 'n' {
|
||||
@@ -1027,10 +1066,11 @@ func (p *parser) scanIntOrFloat(b []byte) (ast.Reference, []byte, error) {
|
||||
return p.builder.Push(ast.Node{
|
||||
Kind: ast.Float,
|
||||
Data: b[:i+3],
|
||||
Raw: p.Range(b[:i+3]),
|
||||
}), b[i+3:], nil
|
||||
}
|
||||
|
||||
return ast.InvalidReference, nil, newDecodeError(b[i:i+1], "unexpected character 'n' while scanning for a number")
|
||||
return ast.InvalidReference, b[i:], newDecodeError(b[i:i+1], "unexpected character 'n' while scanning for a number")
|
||||
}
|
||||
|
||||
break
|
||||
@@ -1049,6 +1089,7 @@ func (p *parser) scanIntOrFloat(b []byte) (ast.Reference, []byte, error) {
|
||||
return p.builder.Push(ast.Node{
|
||||
Kind: kind,
|
||||
Data: b[:i],
|
||||
Raw: p.Range(b[:i]),
|
||||
}), b[i:], nil
|
||||
}
|
||||
|
||||
@@ -1075,11 +1116,11 @@ func isValidBinaryRune(r byte) bool {
|
||||
|
||||
func expect(x byte, b []byte) ([]byte, error) {
|
||||
if len(b) == 0 {
|
||||
return nil, newDecodeError(b, "expected character %c but the document ended here", x)
|
||||
return b, newDecodeError(b, "expected character %c but the document ended here", x)
|
||||
}
|
||||
|
||||
if b[0] != x {
|
||||
return nil, newDecodeError(b[0:1], "expected character %c", x)
|
||||
return b, newDecodeError(b[0:1], "expected character %c", x)
|
||||
}
|
||||
|
||||
return b[1:], nil
|
||||
|
||||
+19
-19
@@ -54,16 +54,16 @@ func scanLiteralString(b []byte) ([]byte, []byte, error) {
|
||||
case '\'':
|
||||
return b[:i+1], b[i+1:], nil
|
||||
case '\n', '\r':
|
||||
return nil, nil, newDecodeError(b[i:i+1], "literal strings cannot have new lines")
|
||||
return nil, b[i+1:], newDecodeError(b[i:i+1], "literal strings cannot have new lines")
|
||||
}
|
||||
size := utf8ValidNext(b[i:])
|
||||
if size == 0 {
|
||||
return nil, nil, newDecodeError(b[i:i+1], "invalid character")
|
||||
return nil, b[i+1:], newDecodeError(b[i:i+1], "invalid character")
|
||||
}
|
||||
i += size
|
||||
}
|
||||
|
||||
return nil, nil, newDecodeError(b[len(b):], "unterminated literal string")
|
||||
return nil, b[len(b):], newDecodeError(b[len(b):], "unterminated literal string")
|
||||
}
|
||||
|
||||
func scanMultilineLiteralString(b []byte) ([]byte, []byte, error) {
|
||||
@@ -98,39 +98,39 @@ func scanMultilineLiteralString(b []byte) ([]byte, []byte, error) {
|
||||
i++
|
||||
|
||||
if i < len(b) && b[i] == '\'' {
|
||||
return nil, nil, newDecodeError(b[i-3:i+1], "''' not allowed in multiline literal string")
|
||||
return nil, b[i:], newDecodeError(b[i-3:i+1], "''' not allowed in multiline literal string")
|
||||
}
|
||||
|
||||
return b[:i], b[i:], nil
|
||||
}
|
||||
case '\r':
|
||||
if len(b) < i+2 {
|
||||
return nil, nil, newDecodeError(b[len(b):], `need a \n after \r`)
|
||||
return nil, b[i:], newDecodeError(b[len(b):], `need a \n after \r`)
|
||||
}
|
||||
if b[i+1] != '\n' {
|
||||
return nil, nil, newDecodeError(b[i:i+2], `need a \n after \r`)
|
||||
return nil, b[i+2:], newDecodeError(b[i:i+2], `need a \n after \r`)
|
||||
}
|
||||
i += 2 // skip the \n
|
||||
continue
|
||||
}
|
||||
size := utf8ValidNext(b[i:])
|
||||
if size == 0 {
|
||||
return nil, nil, newDecodeError(b[i:i+1], "invalid character")
|
||||
return nil, b[i:], newDecodeError(b[i:i+1], "invalid character")
|
||||
}
|
||||
i += size
|
||||
}
|
||||
|
||||
return nil, nil, newDecodeError(b[len(b):], `multiline literal string not terminated by '''`)
|
||||
return nil, b[len(b):], newDecodeError(b[len(b):], `multiline literal string not terminated by '''`)
|
||||
}
|
||||
|
||||
func scanWindowsNewline(b []byte) ([]byte, []byte, error) {
|
||||
const lenCRLF = 2
|
||||
if len(b) < lenCRLF {
|
||||
return nil, nil, newDecodeError(b, "windows new line expected")
|
||||
return nil, b, newDecodeError(b, "windows new line expected")
|
||||
}
|
||||
|
||||
if b[1] != '\n' {
|
||||
return nil, nil, newDecodeError(b, `windows new line should be \r\n`)
|
||||
return nil, b[2:], newDecodeError(b, `windows new line should be \r\n`)
|
||||
}
|
||||
|
||||
return b[:lenCRLF], b[lenCRLF:], nil
|
||||
@@ -169,7 +169,7 @@ func scanComment(b []byte) ([]byte, []byte, error) {
|
||||
}
|
||||
size := utf8ValidNext(b[i:])
|
||||
if size == 0 {
|
||||
return nil, nil, newDecodeError(b[i:i+1], "invalid character in comment")
|
||||
return nil, b[i+1:], newDecodeError(b[i:i+1], "invalid character in comment")
|
||||
}
|
||||
|
||||
i += size
|
||||
@@ -192,17 +192,17 @@ func scanBasicString(b []byte) ([]byte, bool, []byte, error) {
|
||||
case '"':
|
||||
return b[:i+1], escaped, b[i+1:], nil
|
||||
case '\n', '\r':
|
||||
return nil, escaped, nil, newDecodeError(b[i:i+1], "basic strings cannot have new lines")
|
||||
return nil, escaped, b[i+1:], newDecodeError(b[i:i+1], "basic strings cannot have new lines")
|
||||
case '\\':
|
||||
if len(b) < i+2 {
|
||||
return nil, escaped, nil, newDecodeError(b[i:i+1], "need a character after \\")
|
||||
return nil, escaped, b[i+1:], newDecodeError(b[i:i+1], "need a character after \\")
|
||||
}
|
||||
escaped = true
|
||||
i++ // skip the next character
|
||||
}
|
||||
}
|
||||
|
||||
return nil, escaped, nil, newDecodeError(b[len(b):], `basic string not terminated by "`)
|
||||
return nil, escaped, b[len(b):], newDecodeError(b[len(b):], `basic string not terminated by "`)
|
||||
}
|
||||
|
||||
func scanMultilineBasicString(b []byte) ([]byte, bool, []byte, error) {
|
||||
@@ -243,27 +243,27 @@ func scanMultilineBasicString(b []byte) ([]byte, bool, []byte, error) {
|
||||
i++
|
||||
|
||||
if i < len(b) && b[i] == '"' {
|
||||
return nil, escaped, nil, newDecodeError(b[i-3:i+1], `""" not allowed in multiline basic string`)
|
||||
return nil, escaped, b[i+1:], newDecodeError(b[i-3:i+1], `""" not allowed in multiline basic string`)
|
||||
}
|
||||
|
||||
return b[:i], escaped, b[i:], nil
|
||||
}
|
||||
case '\\':
|
||||
if len(b) < i+2 {
|
||||
return nil, escaped, nil, newDecodeError(b[len(b):], "need a character after \\")
|
||||
return nil, escaped, b[len(b):], newDecodeError(b[len(b):], "need a character after \\")
|
||||
}
|
||||
escaped = true
|
||||
i++ // skip the next character
|
||||
case '\r':
|
||||
if len(b) < i+2 {
|
||||
return nil, escaped, nil, newDecodeError(b[len(b):], `need a \n after \r`)
|
||||
return nil, escaped, b[len(b):], newDecodeError(b[len(b):], `need a \n after \r`)
|
||||
}
|
||||
if b[i+1] != '\n' {
|
||||
return nil, escaped, nil, newDecodeError(b[i:i+2], `need a \n after \r`)
|
||||
return nil, escaped, b[i+2:], newDecodeError(b[i:i+2], `need a \n after \r`)
|
||||
}
|
||||
i++ // skip the \n
|
||||
}
|
||||
}
|
||||
|
||||
return nil, escaped, nil, newDecodeError(b[len(b):], `multiline basic string not terminated by """`)
|
||||
return nil, escaped, b[len(b):], newDecodeError(b[len(b):], `multiline basic string not terminated by """`)
|
||||
}
|
||||
|
||||
+67
-29
@@ -79,22 +79,22 @@ func (d *Decoder) DisallowUnknownFields() *Decoder {
|
||||
// strict mode and a field is missing, a `toml.StrictMissingError` is
|
||||
// returned. In any other case, this function returns a standard Go error.
|
||||
//
|
||||
// Type mapping
|
||||
// # Type mapping
|
||||
//
|
||||
// List of supported TOML types and their associated accepted Go types:
|
||||
//
|
||||
// String -> string
|
||||
// Integer -> uint*, int*, depending on size
|
||||
// Float -> float*, depending on size
|
||||
// Boolean -> bool
|
||||
// Offset Date-Time -> time.Time
|
||||
// Local Date-time -> LocalDateTime, time.Time
|
||||
// Local Date -> LocalDate, time.Time
|
||||
// Local Time -> LocalTime, time.Time
|
||||
// Array -> slice and array, depending on elements types
|
||||
// Table -> map and struct
|
||||
// Inline Table -> same as Table
|
||||
// Array of Tables -> same as Array and Table
|
||||
// String -> string
|
||||
// Integer -> uint*, int*, depending on size
|
||||
// Float -> float*, depending on size
|
||||
// Boolean -> bool
|
||||
// Offset Date-Time -> time.Time
|
||||
// Local Date-time -> LocalDateTime, time.Time
|
||||
// Local Date -> LocalDate, time.Time
|
||||
// Local Time -> LocalTime, time.Time
|
||||
// Array -> slice and array, depending on elements types
|
||||
// Table -> map and struct
|
||||
// Inline Table -> same as Table
|
||||
// Array of Tables -> same as Array and Table
|
||||
func (d *Decoder) Decode(v interface{}) error {
|
||||
b, err := ioutil.ReadAll(d.r)
|
||||
if err != nil {
|
||||
@@ -123,7 +123,7 @@ type decoder struct {
|
||||
stashedExpr bool
|
||||
|
||||
// Skip expressions until a table is found. This is set to true when a
|
||||
// table could not be create (missing field in map), so all KV expressions
|
||||
// table could not be created (missing field in map), so all KV expressions
|
||||
// need to be skipped.
|
||||
skipUntilTable bool
|
||||
|
||||
@@ -344,9 +344,9 @@ func (d *decoder) handleArrayTableCollectionLast(key ast.Iterator, v reflect.Val
|
||||
elem := v.Index(idx)
|
||||
_, err := d.handleArrayTable(key, elem)
|
||||
return v, err
|
||||
default:
|
||||
return reflect.Value{}, fmt.Errorf("toml: cannot decode array table into a %s", v.Type())
|
||||
}
|
||||
|
||||
return d.handleArrayTable(key, v)
|
||||
}
|
||||
|
||||
// When parsing an array table expression, each part of the key needs to be
|
||||
@@ -483,7 +483,7 @@ func (d *decoder) handleKeyPart(key ast.Iterator, v reflect.Value, nextFn handle
|
||||
d.errorContext.Struct = t
|
||||
d.errorContext.Field = path
|
||||
|
||||
f := v.FieldByIndex(path)
|
||||
f := fieldByIndex(v, path)
|
||||
x, err := nextFn(key, f)
|
||||
if err != nil || d.skipUntilTable {
|
||||
return reflect.Value{}, err
|
||||
@@ -866,12 +866,27 @@ func (d *decoder) unmarshalFloat(value *ast.Node, v reflect.Value) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func (d *decoder) unmarshalInteger(value *ast.Node, v reflect.Value) error {
|
||||
const (
|
||||
maxInt = int64(^uint(0) >> 1)
|
||||
minInt = -maxInt - 1
|
||||
)
|
||||
const (
|
||||
maxInt = int64(^uint(0) >> 1)
|
||||
minInt = -maxInt - 1
|
||||
)
|
||||
|
||||
// Maximum value of uint for decoding. Currently the decoder parses the integer
|
||||
// into an int64. As a result, on architectures where uint is 64 bits, the
|
||||
// effective maximum uint we can decode is the maximum of int64. On
|
||||
// architectures where uint is 32 bits, the maximum value we can decode is
|
||||
// lower: the maximum of uint32. I didn't find a way to figure out this value at
|
||||
// compile time, so it is computed during initialization.
|
||||
var maxUint int64 = math.MaxInt64
|
||||
|
||||
func init() {
|
||||
m := uint64(^uint(0))
|
||||
if m < uint64(maxUint) {
|
||||
maxUint = int64(m)
|
||||
}
|
||||
}
|
||||
|
||||
func (d *decoder) unmarshalInteger(value *ast.Node, v reflect.Value) error {
|
||||
i, err := parseInteger(value.Data)
|
||||
if err != nil {
|
||||
return err
|
||||
@@ -932,7 +947,7 @@ func (d *decoder) unmarshalInteger(value *ast.Node, v reflect.Value) error {
|
||||
|
||||
r = reflect.ValueOf(uint8(i))
|
||||
case reflect.Uint:
|
||||
if i < 0 {
|
||||
if i < 0 || i > maxUint {
|
||||
return fmt.Errorf("toml: negative number %d does not fit in an uint", i)
|
||||
}
|
||||
|
||||
@@ -1056,7 +1071,7 @@ func (d *decoder) handleKeyValuePart(key ast.Iterator, value *ast.Node, v reflec
|
||||
d.errorContext.Struct = t
|
||||
d.errorContext.Field = path
|
||||
|
||||
f := v.FieldByIndex(path)
|
||||
f := fieldByIndex(v, path)
|
||||
x, err := d.handleKeyValueInner(key, value, f)
|
||||
if err != nil {
|
||||
return reflect.Value{}, err
|
||||
@@ -1120,6 +1135,21 @@ func initAndDereferencePointer(v reflect.Value) reflect.Value {
|
||||
return elem
|
||||
}
|
||||
|
||||
// Same as reflect.Value.FieldByIndex, but creates pointers if needed.
|
||||
func fieldByIndex(v reflect.Value, path []int) reflect.Value {
|
||||
for i, x := range path {
|
||||
v = v.Field(x)
|
||||
|
||||
if i < len(path)-1 && v.Kind() == reflect.Pointer {
|
||||
if v.IsNil() {
|
||||
v.Set(reflect.New(v.Type().Elem()))
|
||||
}
|
||||
v = v.Elem()
|
||||
}
|
||||
}
|
||||
return v
|
||||
}
|
||||
|
||||
type fieldPathsMap = map[string][]int
|
||||
|
||||
var globalFieldPathsCache atomic.Value // map[danger.TypeID]fieldPathsMap
|
||||
@@ -1167,11 +1197,6 @@ func forEachField(t reflect.Type, path []int, do func(name string, path []int))
|
||||
fieldPath := append(path, i)
|
||||
fieldPath = fieldPath[:len(fieldPath):len(fieldPath)]
|
||||
|
||||
if f.Anonymous {
|
||||
forEachField(f.Type, fieldPath, do)
|
||||
continue
|
||||
}
|
||||
|
||||
name := f.Tag.Get("toml")
|
||||
if name == "-" {
|
||||
continue
|
||||
@@ -1180,6 +1205,19 @@ func forEachField(t reflect.Type, path []int, do func(name string, path []int))
|
||||
if i := strings.IndexByte(name, ','); i >= 0 {
|
||||
name = name[:i]
|
||||
}
|
||||
|
||||
if f.Anonymous && name == "" {
|
||||
t2 := f.Type
|
||||
if t2.Kind() == reflect.Pointer {
|
||||
t2 = t2.Elem()
|
||||
}
|
||||
|
||||
if t2.Kind() == reflect.Struct {
|
||||
forEachField(t2, fieldPath, do)
|
||||
}
|
||||
continue
|
||||
}
|
||||
|
||||
if name == "" {
|
||||
name = f.Name
|
||||
}
|
||||
|
||||
+149
-20
@@ -554,7 +554,7 @@ wibble = 'wobble'
|
||||
[foo]
|
||||
|
||||
[foo.bar]
|
||||
huey = 'dewey'
|
||||
huey = 'dewey'
|
||||
`,
|
||||
gen: func() test {
|
||||
m := map[string]interface{}{}
|
||||
@@ -1735,6 +1735,28 @@ B = "data"`,
|
||||
}
|
||||
},
|
||||
},
|
||||
{
|
||||
desc: "kv that points to a slice",
|
||||
input: "a.b.c = 'foo'",
|
||||
gen: func() test {
|
||||
doc := map[string][]string{}
|
||||
return test{
|
||||
target: &doc,
|
||||
err: true,
|
||||
}
|
||||
},
|
||||
},
|
||||
{
|
||||
desc: "kv that points to a pointer to a slice",
|
||||
input: "a.b.c = 'foo'",
|
||||
gen: func() test {
|
||||
doc := map[string]*[]string{}
|
||||
return test{
|
||||
target: &doc,
|
||||
err: true,
|
||||
}
|
||||
},
|
||||
},
|
||||
}
|
||||
|
||||
for _, e := range examples {
|
||||
@@ -1876,8 +1898,7 @@ key2 = "missing2"
|
||||
key3 = "missing3"
|
||||
key4 = "value4"
|
||||
`,
|
||||
expected: `
|
||||
2| key1 = "value1"
|
||||
expected: `2| key1 = "value1"
|
||||
3| key2 = "missing2"
|
||||
| ~~~~ missing field
|
||||
4| key3 = "missing3"
|
||||
@@ -1887,8 +1908,7 @@ key4 = "value4"
|
||||
3| key2 = "missing2"
|
||||
4| key3 = "missing3"
|
||||
| ~~~~ missing field
|
||||
5| key4 = "value4"
|
||||
`,
|
||||
5| key4 = "value4"`,
|
||||
target: &struct {
|
||||
Key1 string
|
||||
Key4 string
|
||||
@@ -1897,10 +1917,8 @@ key4 = "value4"
|
||||
{
|
||||
desc: "multi-part key",
|
||||
input: `a.short.key="foo"`,
|
||||
expected: `
|
||||
1| a.short.key="foo"
|
||||
| ~~~~~~~~~~~ missing field
|
||||
`,
|
||||
expected: `1| a.short.key="foo"
|
||||
| ~~~~~~~~~~~ missing field`,
|
||||
},
|
||||
{
|
||||
desc: "missing table",
|
||||
@@ -1908,24 +1926,19 @@ key4 = "value4"
|
||||
[foo]
|
||||
bar = 42
|
||||
`,
|
||||
expected: `
|
||||
2| [foo]
|
||||
expected: `2| [foo]
|
||||
| ~~~ missing table
|
||||
3| bar = 42
|
||||
`,
|
||||
3| bar = 42`,
|
||||
},
|
||||
|
||||
{
|
||||
desc: "missing array table",
|
||||
input: `
|
||||
[[foo]]
|
||||
bar = 42
|
||||
`,
|
||||
expected: `
|
||||
2| [[foo]]
|
||||
bar = 42`,
|
||||
expected: `2| [[foo]]
|
||||
| ~~~ missing table
|
||||
3| bar = 42
|
||||
`,
|
||||
3| bar = 42`,
|
||||
},
|
||||
}
|
||||
|
||||
@@ -1944,7 +1957,7 @@ bar = 42
|
||||
|
||||
var tsm *toml.StrictMissingError
|
||||
if errors.As(err, &tsm) {
|
||||
equalStringsIgnoreNewlines(t, e.expected, tsm.String())
|
||||
assert.Equal(t, e.expected, tsm.String())
|
||||
} else {
|
||||
t.Fatalf("err should have been a *toml.StrictMissingError, but got %s (%T)", err, err)
|
||||
}
|
||||
@@ -2380,6 +2393,79 @@ func TestIssue714(t *testing.T) {
|
||||
require.Error(t, err)
|
||||
}
|
||||
|
||||
func TestIssue772(t *testing.T) {
|
||||
type FileHandling struct {
|
||||
FilePattern string `toml:"pattern"`
|
||||
}
|
||||
|
||||
type Config struct {
|
||||
FileHandling `toml:"filehandling"`
|
||||
}
|
||||
|
||||
var defaultConfigFile = []byte(`
|
||||
[filehandling]
|
||||
pattern = "reach-masterdev-"`)
|
||||
|
||||
config := Config{}
|
||||
err := toml.Unmarshal(defaultConfigFile, &config)
|
||||
require.NoError(t, err)
|
||||
require.Equal(t, "reach-masterdev-", config.FileHandling.FilePattern)
|
||||
}
|
||||
|
||||
func TestIssue774(t *testing.T) {
|
||||
type ScpData struct {
|
||||
Host string `json:"host"`
|
||||
}
|
||||
|
||||
type GenConfig struct {
|
||||
SCP []ScpData `toml:"scp" comment:"Array of Secure Copy Configurations"`
|
||||
}
|
||||
|
||||
c := &GenConfig{}
|
||||
c.SCP = []ScpData{{Host: "main.domain.com"}}
|
||||
|
||||
b, err := toml.Marshal(c)
|
||||
require.NoError(t, err)
|
||||
|
||||
expected := `# Array of Secure Copy Configurations
|
||||
[[scp]]
|
||||
Host = 'main.domain.com'
|
||||
`
|
||||
|
||||
require.Equal(t, expected, string(b))
|
||||
}
|
||||
|
||||
func TestIssue799(t *testing.T) {
|
||||
const testTOML = `
|
||||
# notice the double brackets
|
||||
[[test]]
|
||||
answer = 42
|
||||
`
|
||||
|
||||
var s struct {
|
||||
// should be []map[string]int
|
||||
Test map[string]int `toml:"test"`
|
||||
}
|
||||
|
||||
err := toml.Unmarshal([]byte(testTOML), &s)
|
||||
require.Error(t, err)
|
||||
}
|
||||
|
||||
func TestIssue807(t *testing.T) {
|
||||
type A struct {
|
||||
Name string `toml:"name"`
|
||||
}
|
||||
|
||||
type M struct {
|
||||
*A
|
||||
}
|
||||
|
||||
var m M
|
||||
err := toml.Unmarshal([]byte(`name = 'foo'`), &m)
|
||||
require.NoError(t, err)
|
||||
require.Equal(t, "foo", m.Name)
|
||||
}
|
||||
|
||||
func TestUnmarshalDecodeErrors(t *testing.T) {
|
||||
examples := []struct {
|
||||
desc string
|
||||
@@ -2831,6 +2917,36 @@ world'`,
|
||||
}
|
||||
}
|
||||
|
||||
func TestOmitEmpty(t *testing.T) {
|
||||
type inner struct {
|
||||
private string
|
||||
Skip string `toml:"-"`
|
||||
V string
|
||||
}
|
||||
|
||||
type elem struct {
|
||||
Foo string `toml:",omitempty"`
|
||||
Bar string `toml:",omitempty"`
|
||||
Inner inner `toml:",omitempty"`
|
||||
}
|
||||
|
||||
type doc struct {
|
||||
X []elem `toml:",inline"`
|
||||
}
|
||||
|
||||
d := doc{X: []elem{elem{
|
||||
Foo: "test",
|
||||
Inner: inner{
|
||||
V: "alue",
|
||||
},
|
||||
}}}
|
||||
|
||||
b, err := toml.Marshal(d)
|
||||
require.NoError(t, err)
|
||||
|
||||
require.Equal(t, "X = [{Foo = 'test', Inner = {V = 'alue'}}]\n", string(b))
|
||||
}
|
||||
|
||||
func TestUnmarshalTags(t *testing.T) {
|
||||
type doc struct {
|
||||
Dash string `toml:"-,"`
|
||||
@@ -3172,3 +3288,16 @@ func TestUnmarshal_RecursiveTableArray(t *testing.T) {
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestUnmarshalEmbedNonString(t *testing.T) {
|
||||
type Foo []byte
|
||||
type doc struct {
|
||||
Foo
|
||||
}
|
||||
|
||||
d := doc{}
|
||||
|
||||
err := toml.Unmarshal([]byte(`foo = 'bar'`), &d)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, d.Foo)
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user