Fix memory leak in tokenize_file, per report from Vadim Passynkov.

This commit is contained in:
Tom Lane 2004-09-18 01:23:12 +00:00
parent 50ff806700
commit f44a64901e

View File

@ -10,7 +10,7 @@
* *
* *
* IDENTIFICATION * IDENTIFICATION
* $Header: /cvsroot/pgsql/src/backend/libpq/hba.c,v 1.116.2.2 2004/05/25 19:11:26 tgl Exp $ * $Header: /cvsroot/pgsql/src/backend/libpq/hba.c,v 1.116.2.3 2004/09/18 01:23:12 tgl Exp $
* *
*------------------------------------------------------------------------- *-------------------------------------------------------------------------
*/ */
@ -168,6 +168,9 @@ next_token(FILE *fp, char *buf, const int bufsz)
* Tokenize file and handle file inclusion and comma lists. We have * Tokenize file and handle file inclusion and comma lists. We have
* to break apart the commas to expand any file names then * to break apart the commas to expand any file names then
* reconstruct with commas. * reconstruct with commas.
*
* The result is always a palloc'd string. If it's zero-length then
* we have reached EOL.
*/ */
static char * static char *
next_token_expand(FILE *file) next_token_expand(FILE *file)
@ -333,6 +336,8 @@ tokenize_file(FILE *file)
{ {
/* we are at real or logical EOL, so force a new line List */ /* we are at real or logical EOL, so force a new line List */
next_line = NIL; next_line = NIL;
/* Don't forget to pfree the next_token_expand result */
pfree(buf);
} }
/* Advance line number whenever we reach EOL */ /* Advance line number whenever we reach EOL */