1// Copyright 2011 The Go Authors. All rights reserved. 2// Use of this source code is governed by a BSD-style 3// license that can be found in the LICENSE file. 4 5// Package csv reads and writes comma-separated values (CSV) files. 6// There are many kinds of CSV files; this package supports the format 7// described in RFC 4180, except that [Writer] uses LF 8// instead of CRLF as newline character by default. 9// 10// A csv file contains zero or more records of one or more fields per record. 11// Each record is separated by the newline character. The final record may 12// optionally be followed by a newline character. 13// 14// field1,field2,field3 15// 16// White space is considered part of a field. 17// 18// Carriage returns before newline characters are silently removed. 19// 20// Blank lines are ignored. A line with only whitespace characters (excluding 21// the ending newline character) is not considered a blank line. 22// 23// Fields which start and stop with the quote character " are called 24// quoted-fields. The beginning and ending quote are not part of the 25// field. 26// 27// The source: 28// 29// normal string,"quoted-field" 30// 31// results in the fields 32// 33// {`normal string`, `quoted-field`} 34// 35// Within a quoted-field a quote character followed by a second quote 36// character is considered a single quote. 37// 38// "the ""word"" is true","a ""quoted-field""" 39// 40// results in 41// 42// {`the "word" is true`, `a "quoted-field"`} 43// 44// Newlines and commas may be included in a quoted-field 45// 46// "Multi-line 47// field","comma is ," 48// 49// results in 50// 51// {`Multi-line 52// field`, `comma is ,`} 53package csv 54 55import ( 56 "bufio" 57 "bytes" 58 "errors" 59 "fmt" 60 "io" 61 "unicode" 62 "unicode/utf8" 63) 64 65// A ParseError is returned for parsing errors. 66// Line and column numbers are 1-indexed. 67type ParseError struct { 68 StartLine int // Line where the record starts 69 Line int // Line where the error occurred 70 Column int // Column (1-based byte index) where the error occurred 71 Err error // The actual error 72} 73 74func (e *ParseError) Error() string { 75 if e.Err == ErrFieldCount { 76 return fmt.Sprintf("record on line %d: %v", e.Line, e.Err) 77 } 78 if e.StartLine != e.Line { 79 return fmt.Sprintf("record on line %d; parse error on line %d, column %d: %v", e.StartLine, e.Line, e.Column, e.Err) 80 } 81 return fmt.Sprintf("parse error on line %d, column %d: %v", e.Line, e.Column, e.Err) 82} 83 84func (e *ParseError) Unwrap() error { return e.Err } 85 86// These are the errors that can be returned in [ParseError.Err]. 87var ( 88 ErrBareQuote = errors.New("bare \" in non-quoted-field") 89 ErrQuote = errors.New("extraneous or missing \" in quoted-field") 90 ErrFieldCount = errors.New("wrong number of fields") 91 92 // Deprecated: ErrTrailingComma is no longer used. 93 ErrTrailingComma = errors.New("extra delimiter at end of line") 94) 95 96var errInvalidDelim = errors.New("csv: invalid field or comment delimiter") 97 98func validDelim(r rune) bool { 99 return r != 0 && r != '"' && r != '\r' && r != '\n' && utf8.ValidRune(r) && r != utf8.RuneError 100} 101 102// A Reader reads records from a CSV-encoded file. 103// 104// As returned by [NewReader], a Reader expects input conforming to RFC 4180. 105// The exported fields can be changed to customize the details before the 106// first call to [Reader.Read] or [Reader.ReadAll]. 107// 108// The Reader converts all \r\n sequences in its input to plain \n, 109// including in multiline field values, so that the returned data does 110// not depend on which line-ending convention an input file uses. 111type Reader struct { 112 // Comma is the field delimiter. 113 // It is set to comma (',') by NewReader. 114 // Comma must be a valid rune and must not be \r, \n, 115 // or the Unicode replacement character (0xFFFD). 116 Comma rune 117 118 // Comment, if not 0, is the comment character. Lines beginning with the 119 // Comment character without preceding whitespace are ignored. 120 // With leading whitespace the Comment character becomes part of the 121 // field, even if TrimLeadingSpace is true. 122 // Comment must be a valid rune and must not be \r, \n, 123 // or the Unicode replacement character (0xFFFD). 124 // It must also not be equal to Comma. 125 Comment rune 126 127 // FieldsPerRecord is the number of expected fields per record. 128 // If FieldsPerRecord is positive, Read requires each record to 129 // have the given number of fields. If FieldsPerRecord is 0, Read sets it to 130 // the number of fields in the first record, so that future records must 131 // have the same field count. If FieldsPerRecord is negative, no check is 132 // made and records may have a variable number of fields. 133 FieldsPerRecord int 134 135 // If LazyQuotes is true, a quote may appear in an unquoted field and a 136 // non-doubled quote may appear in a quoted field. 137 LazyQuotes bool 138 139 // If TrimLeadingSpace is true, leading white space in a field is ignored. 140 // This is done even if the field delimiter, Comma, is white space. 141 TrimLeadingSpace bool 142 143 // ReuseRecord controls whether calls to Read may return a slice sharing 144 // the backing array of the previous call's returned slice for performance. 145 // By default, each call to Read returns newly allocated memory owned by the caller. 146 ReuseRecord bool 147 148 // Deprecated: TrailingComma is no longer used. 149 TrailingComma bool 150 151 r *bufio.Reader 152 153 // numLine is the current line being read in the CSV file. 154 numLine int 155 156 // offset is the input stream byte offset of the current reader position. 157 offset int64 158 159 // rawBuffer is a line buffer only used by the readLine method. 160 rawBuffer []byte 161 162 // recordBuffer holds the unescaped fields, one after another. 163 // The fields can be accessed by using the indexes in fieldIndexes. 164 // E.g., For the row `a,"b","c""d",e`, recordBuffer will contain `abc"de` 165 // and fieldIndexes will contain the indexes [1, 2, 5, 6]. 166 recordBuffer []byte 167 168 // fieldIndexes is an index of fields inside recordBuffer. 169 // The i'th field ends at offset fieldIndexes[i] in recordBuffer. 170 fieldIndexes []int 171 172 // fieldPositions is an index of field positions for the 173 // last record returned by Read. 174 fieldPositions []position 175 176 // lastRecord is a record cache and only used when ReuseRecord == true. 177 lastRecord []string 178} 179 180// NewReader returns a new Reader that reads from r. 181func NewReader(r io.Reader) *Reader { 182 return &Reader{ 183 Comma: ',', 184 r: bufio.NewReader(r), 185 } 186} 187 188// Read reads one record (a slice of fields) from r. 189// If the record has an unexpected number of fields, 190// Read returns the record along with the error [ErrFieldCount]. 191// If the record contains a field that cannot be parsed, 192// Read returns a partial record along with the parse error. 193// The partial record contains all fields read before the error. 194// If there is no data left to be read, Read returns nil, [io.EOF]. 195// If [Reader.ReuseRecord] is true, the returned slice may be shared 196// between multiple calls to Read. 197func (r *Reader) Read() (record []string, err error) { 198 if r.ReuseRecord { 199 record, err = r.readRecord(r.lastRecord) 200 r.lastRecord = record 201 } else { 202 record, err = r.readRecord(nil) 203 } 204 return record, err 205} 206 207// FieldPos returns the line and column corresponding to 208// the start of the field with the given index in the slice most recently 209// returned by [Reader.Read]. Numbering of lines and columns starts at 1; 210// columns are counted in bytes, not runes. 211// 212// If this is called with an out-of-bounds index, it panics. 213func (r *Reader) FieldPos(field int) (line, column int) { 214 if field < 0 || field >= len(r.fieldPositions) { 215 panic("out of range index passed to FieldPos") 216 } 217 p := &r.fieldPositions[field] 218 return p.line, p.col 219} 220 221// InputOffset returns the input stream byte offset of the current reader 222// position. The offset gives the location of the end of the most recently 223// read row and the beginning of the next row. 224func (r *Reader) InputOffset() int64 { 225 return r.offset 226} 227 228// pos holds the position of a field in the current line. 229type position struct { 230 line, col int 231} 232 233// ReadAll reads all the remaining records from r. 234// Each record is a slice of fields. 235// A successful call returns err == nil, not err == [io.EOF]. Because ReadAll is 236// defined to read until EOF, it does not treat end of file as an error to be 237// reported. 238func (r *Reader) ReadAll() (records [][]string, err error) { 239 for { 240 record, err := r.readRecord(nil) 241 if err == io.EOF { 242 return records, nil 243 } 244 if err != nil { 245 return nil, err 246 } 247 records = append(records, record) 248 } 249} 250 251// readLine reads the next line (with the trailing endline). 252// If EOF is hit without a trailing endline, it will be omitted. 253// If some bytes were read, then the error is never [io.EOF]. 254// The result is only valid until the next call to readLine. 255func (r *Reader) readLine() ([]byte, error) { 256 line, err := r.r.ReadSlice('\n') 257 if err == bufio.ErrBufferFull { 258 r.rawBuffer = append(r.rawBuffer[:0], line...) 259 for err == bufio.ErrBufferFull { 260 line, err = r.r.ReadSlice('\n') 261 r.rawBuffer = append(r.rawBuffer, line...) 262 } 263 line = r.rawBuffer 264 } 265 readSize := len(line) 266 if readSize > 0 && err == io.EOF { 267 err = nil 268 // For backwards compatibility, drop trailing \r before EOF. 269 if line[readSize-1] == '\r' { 270 line = line[:readSize-1] 271 } 272 } 273 r.numLine++ 274 r.offset += int64(readSize) 275 // Normalize \r\n to \n on all input lines. 276 if n := len(line); n >= 2 && line[n-2] == '\r' && line[n-1] == '\n' { 277 line[n-2] = '\n' 278 line = line[:n-1] 279 } 280 return line, err 281} 282 283// lengthNL reports the number of bytes for the trailing \n. 284func lengthNL(b []byte) int { 285 if len(b) > 0 && b[len(b)-1] == '\n' { 286 return 1 287 } 288 return 0 289} 290 291// nextRune returns the next rune in b or utf8.RuneError. 292func nextRune(b []byte) rune { 293 r, _ := utf8.DecodeRune(b) 294 return r 295} 296 297func (r *Reader) readRecord(dst []string) ([]string, error) { 298 if r.Comma == r.Comment || !validDelim(r.Comma) || (r.Comment != 0 && !validDelim(r.Comment)) { 299 return nil, errInvalidDelim 300 } 301 302 // Read line (automatically skipping past empty lines and any comments). 303 var line []byte 304 var errRead error 305 for errRead == nil { 306 line, errRead = r.readLine() 307 if r.Comment != 0 && nextRune(line) == r.Comment { 308 line = nil 309 continue // Skip comment lines 310 } 311 if errRead == nil && len(line) == lengthNL(line) { 312 line = nil 313 continue // Skip empty lines 314 } 315 break 316 } 317 if errRead == io.EOF { 318 return nil, errRead 319 } 320 321 // Parse each field in the record. 322 var err error 323 const quoteLen = len(`"`) 324 commaLen := utf8.RuneLen(r.Comma) 325 recLine := r.numLine // Starting line for record 326 r.recordBuffer = r.recordBuffer[:0] 327 r.fieldIndexes = r.fieldIndexes[:0] 328 r.fieldPositions = r.fieldPositions[:0] 329 pos := position{line: r.numLine, col: 1} 330parseField: 331 for { 332 if r.TrimLeadingSpace { 333 i := bytes.IndexFunc(line, func(r rune) bool { 334 return !unicode.IsSpace(r) 335 }) 336 if i < 0 { 337 i = len(line) 338 pos.col -= lengthNL(line) 339 } 340 line = line[i:] 341 pos.col += i 342 } 343 if len(line) == 0 || line[0] != '"' { 344 // Non-quoted string field 345 i := bytes.IndexRune(line, r.Comma) 346 field := line 347 if i >= 0 { 348 field = field[:i] 349 } else { 350 field = field[:len(field)-lengthNL(field)] 351 } 352 // Check to make sure a quote does not appear in field. 353 if !r.LazyQuotes { 354 if j := bytes.IndexByte(field, '"'); j >= 0 { 355 col := pos.col + j 356 err = &ParseError{StartLine: recLine, Line: r.numLine, Column: col, Err: ErrBareQuote} 357 break parseField 358 } 359 } 360 r.recordBuffer = append(r.recordBuffer, field...) 361 r.fieldIndexes = append(r.fieldIndexes, len(r.recordBuffer)) 362 r.fieldPositions = append(r.fieldPositions, pos) 363 if i >= 0 { 364 line = line[i+commaLen:] 365 pos.col += i + commaLen 366 continue parseField 367 } 368 break parseField 369 } else { 370 // Quoted string field 371 fieldPos := pos 372 line = line[quoteLen:] 373 pos.col += quoteLen 374 for { 375 i := bytes.IndexByte(line, '"') 376 if i >= 0 { 377 // Hit next quote. 378 r.recordBuffer = append(r.recordBuffer, line[:i]...) 379 line = line[i+quoteLen:] 380 pos.col += i + quoteLen 381 switch rn := nextRune(line); { 382 case rn == '"': 383 // `""` sequence (append quote). 384 r.recordBuffer = append(r.recordBuffer, '"') 385 line = line[quoteLen:] 386 pos.col += quoteLen 387 case rn == r.Comma: 388 // `",` sequence (end of field). 389 line = line[commaLen:] 390 pos.col += commaLen 391 r.fieldIndexes = append(r.fieldIndexes, len(r.recordBuffer)) 392 r.fieldPositions = append(r.fieldPositions, fieldPos) 393 continue parseField 394 case lengthNL(line) == len(line): 395 // `"\n` sequence (end of line). 396 r.fieldIndexes = append(r.fieldIndexes, len(r.recordBuffer)) 397 r.fieldPositions = append(r.fieldPositions, fieldPos) 398 break parseField 399 case r.LazyQuotes: 400 // `"` sequence (bare quote). 401 r.recordBuffer = append(r.recordBuffer, '"') 402 default: 403 // `"*` sequence (invalid non-escaped quote). 404 err = &ParseError{StartLine: recLine, Line: r.numLine, Column: pos.col - quoteLen, Err: ErrQuote} 405 break parseField 406 } 407 } else if len(line) > 0 { 408 // Hit end of line (copy all data so far). 409 r.recordBuffer = append(r.recordBuffer, line...) 410 if errRead != nil { 411 break parseField 412 } 413 pos.col += len(line) 414 line, errRead = r.readLine() 415 if len(line) > 0 { 416 pos.line++ 417 pos.col = 1 418 } 419 if errRead == io.EOF { 420 errRead = nil 421 } 422 } else { 423 // Abrupt end of file (EOF or error). 424 if !r.LazyQuotes && errRead == nil { 425 err = &ParseError{StartLine: recLine, Line: pos.line, Column: pos.col, Err: ErrQuote} 426 break parseField 427 } 428 r.fieldIndexes = append(r.fieldIndexes, len(r.recordBuffer)) 429 r.fieldPositions = append(r.fieldPositions, fieldPos) 430 break parseField 431 } 432 } 433 } 434 } 435 if err == nil { 436 err = errRead 437 } 438 439 // Create a single string and create slices out of it. 440 // This pins the memory of the fields together, but allocates once. 441 str := string(r.recordBuffer) // Convert to string once to batch allocations 442 dst = dst[:0] 443 if cap(dst) < len(r.fieldIndexes) { 444 dst = make([]string, len(r.fieldIndexes)) 445 } 446 dst = dst[:len(r.fieldIndexes)] 447 var preIdx int 448 for i, idx := range r.fieldIndexes { 449 dst[i] = str[preIdx:idx] 450 preIdx = idx 451 } 452 453 // Check or update the expected fields per record. 454 if r.FieldsPerRecord > 0 { 455 if len(dst) != r.FieldsPerRecord && err == nil { 456 err = &ParseError{ 457 StartLine: recLine, 458 Line: recLine, 459 Column: 1, 460 Err: ErrFieldCount, 461 } 462 } 463 } else if r.FieldsPerRecord == 0 { 464 r.FieldsPerRecord = len(dst) 465 } 466 return dst, err 467} 468