SARG Schedule Starting Multiple force refreshs at scheduled time



  • Hi,

    I seem to have a problem with SARG strating multiple force refreshes at the schedules time. See log extract below:

    Jan 15 16:00:03	php: : Sarg: force refresh now with -d `date +%d/%m/%Y`-`date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 15 16:00:03	php: : Sarg: force refresh now with -d `date +%d/%m/%Y`-`date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 15 16:00:03	php: : Sarg: force refresh now with -d `date +%d/%m/%Y`-`date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 15 16:00:02	php: : Sarg: force refresh now with -d `date +%d/%m/%Y`-`date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 15 16:00:02	php: : Sarg: force refresh now with -d `date +%d/%m/%Y`-`date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 15 16:00:02	php: : Sarg: force refresh now with -d `date +%d/%m/%Y`-`date +%d/%m/%Y` args, compress(on) and none action after sarg finish
    

    From this I take it that there is a CRON job setup to run the same command 6 times. When I look in Services->CRON I only see one entry. Why then is the command run 6 times? Is there another place where I can check this?

    Thanks



  • If it helps I have noticed that the number of times the command is run depends if I enable or disable schedules within SARG. i.e. If I enable a new schedule, then disable it for whatever reason an additional instance is still trying to run.

    What file are the SARG schedules saved in? I'm thinking there may be a bug where the disabled schedule is not fully removed from it.



  • Anybody have any thoughts on this? Why would sarg report refresh appear to start so many times in the log file?



  • Could you give us more of the log file. As well what is your Sarg conf?

    What does it say in your crontab anyway?

    Do you have this set in your sarg.conf?

    # TAG: overwrite_report yes|no
    #      yes - if report date already exist then will be overwrited.
    #       no - if report date already exist then will be renamed to filename.n, filename.n+1
    #
    overwrite_report yes
    

    If so do you actually see multiple reports with filename.n+1?



  • Sure no problem. Just a note, when I manually do a force refresh everything works just fine.

    sarg.conf:

    # sarg.conf
    #
    # TAG:  access_log file
    #       Where is the access.log
    #       sarg -l file
    #
    access_log /var/log/dansguardian/access.log
    
    # TAG: graphs yes|no
    #	Use graphics where is possible.
    #           graph_days_bytes_bar_color blue|green|yellow|orange|brown|red
    #
    graphs no
    #graph_days_bytes_bar_color orange
    
    # TAG:  graph_font
    #       The full path to the TTF font file to use to create the graphs. It is required
    #       if graphs is set to yes.
    #
    #graph_font /usr/share/fonts/truetype/ttf-dejavu/DejaVuSans.ttf
    
    # TAG:	title
    # 	Especify the title for html page.
    #
    #title "Squid User Access Reports"
    
    # TAG:	font_face
    # 	Especify the font for html page.
    #
    #font_face Tahoma,Verdana,Arial
    
    # TAG:	header_color
    # 	Especify the header color
    #
    #header_color darkblue
    
    # TAG:	header_bgcolor
    # 	Especify the header bgcolor
    #
    #header_bgcolor blanchedalmond
    
    # TAG:	font_size
    # 	Especify the text font size
    #
    #font_size 9px
    
    # TAG:	header_font_size
    # 	Especify the header font size
    #
    #header_font_size 9px
    
    # TAG:	title_font_size
    # 	Especify the title font size
    #
    #title_font_size 11px
    
    # TAG:	background_color
    # TAG:	background_color
    #	Html page background color
    #
    # background_color white
    
    # TAG:	text_color
    #	Html page text color
    #
    #text_color #000000
    
    # TAG:	text_bgcolor
    #	Html page text background color
    #
    #text_bgcolor lavender
    
    # TAG:	title_color
    #	Html page title color
    #
    #title_color green
    
    # TAG:	logo_image
    #	Html page logo.
    #
    #logo_image none
    
    # TAG:	logo_text
    #	Html page logo text.
    #
    #logo_text ""
    
    # TAG:	logo_text_color
    #	Html page logo texti color.
    #
    #logo_text_color #000000
    
    # TAG:	logo_image_size
    #	Html page logo image size.
    #       width height
    #
    #image_size 80 45
    
    # TAG:	background_image
    #	Html page background image
    #
    #background_image none
    
    # TAG:  password
    #       User password file used by Squid authentication scheme
    #       If used, generate reports just for that users.
    #
    #password none
    
    # TAG:  temporary_dir
    #       Temporary directory name for work files
    #       sarg -w dir
    #
    #temporary_dir /tmp
    
    # TAG:  output_dir
    #       The reports will be saved in that directory
    #       sarg -o dir
    #
    output_dir /usr/local/sarg-reports
    
    # TAG:  anonymous_output_files yes/no
    #       Use anonymous file and directory names in the report. If it is set to
    #       no (the default), the user id/ip/name is slightly mangled to create a
    #       suitable file name to store the report of the user but the user's
    #       identity can easily be guessed from the mangled name. If this option is
    #       set, any file or directory belonging to the user is replaced by a short
    #       number.  The purpose is to hide the identity of the user when looking
    #       at the report file names but it may serve to shorten the path too.
    #
    anonymous_output_files no
    
    # TAG:  output_email
    #       Email address to send the reports. If you use this tag, no html reports will be generated.
    #       sarg -e email
    #
    #output_email none
    
    # TAG:  resolve_ip yes/no
    #       Convert ip address to dns name
    #       sarg -n
    resolve_ip yes
    
    # TAG:  user_ip yes/no
    #       Use Ip Address instead userid in reports.
    #       sarg -p
    user_ip yes
    
    # TAG:  topuser_sort_field field normal/reverse
    #       Sort field for the Topuser Report.
    #       Allowed fields: USER CONNECT BYTES TIME
    #
    topuser_sort_field BYTES reverse
    
    # TAG:  user_sort_field field normal/reverse
    #       Sort field for the User Report.
    #       Allowed fields: SITE CONNECT BYTES TIME
    #
    user_sort_field BYTES reverse
    
    # TAG:  exclude_users file
    #       users within the file will be excluded from reports.
    #       you can use indexonly to have only index.html file.
    #
    exclude_users /usr/local/etc/sarg/exclude_users.conf
    
    # TAG:  exclude_hosts file
    #       Hosts, domains or subnets will be excluded from reports.
    #
    #       Eg.: 192.168.10.10   - exclude ip address only
    #            192.168.10.0/24 - exclude full C class
    #            s1.acme.foo     - exclude hostname only
    #            *.acme.foo      - exclude full domain name
    #
    exclude_hosts /usr/local/etc/sarg/exclude_hosts.conf
    
    # TAG:  useragent_log file
    #       useragent.log file patch to generate useragent report.
    #
    #useragent_log none
    
    # TAG:  date_format
    #       Date format in reports: e (European=dd/mm/yy), u (American=mm/dd/yy), w (Weekly=yy.ww)
    #
    #date_format u
    date_format u
    
    # TAG:  per_user_limit file MB
    #       Saves userid on file if download exceed n MB.
    #       This option allow you to disable user access if user exceed a download limit.
    #
    #per_user_limit none
    
    # TAG: lastlog n
    #      How many reports files must be kept in reports directory.
    #      The oldest report file will be automatically removed.
    #      0 - no limit.
    #
    #lastlog 0
    lastlog 14
    
    # TAG: remove_temp_files yes
    #      Remove temporary files: geral, usuarios, top, periodo from root report directory.
    #
    remove_temp_files yes
    
    # TAG: index yes|no|only
    #      Generate the main index.html.
    #      only - generate only the main index.html
    #
    index yes
    
    # TAG: index_tree date|file
    #      How to generate the index.
    #
    index_tree file
    
    # TAG: index_fields
    #      The columns to show in the index of the reports
    #      Columns are: dirsize
    #
    #index_fields dirsize
    
    # TAG: overwrite_report yes|no
    #      yes - if report date already exist then will be overwrited.
    #       no - if report date already exist then will be renamed to filename.n, filename.n+1
    #
    overwrite_report yes
    
    # TAG: records_without_userid ignore|ip|everybody
    #      What can I do with records without user id (no authentication) in access.log file ?
    #
    #      ignore - This record will be ignored.
    #          ip - Use ip address instead. (default)
    #   everybody - Use "everybody" instead.
    #
    #records_without_userid ip
    
    # TAG: use_comma no|yes
    #      Use comma instead point in reports.
    #      Eg.: use_comma yes => 23,450,110
    #           use_comma no  => 23.450.110
    #
    use_comma yes
    
    # TAG: mail_utility
    #      Mail command to use to send reports via SMTP. Sarg calls it like this:
    #         mail_utility -s "SARG report, date" "output_email" <"mail_content"
    #
    #      Therefore, it is possible to add more arguments to the command by specifying them
    #      here.
    #
    #      If you need too, you can use a shell script to process the content of /dev/stdin
    #      (/dev/stdin is the mail_content passed by sarg to the script) and call whatever
    #      command you like. It is not limited to mailing the report via SMTP.
    #
    #      Don't forget to quote the command if necessary (i.e. if the path contains
    #      characters that must be quoted).
    #
    #mail_utility mailx
    
    # TAG: topsites_num n
    #      How many sites in topsites report.
    #
    #topsites_num 100
    
    # TAG: topsites_sort_order CONNECT|BYTES|TIME A|D
    #      Sort for topsites report, where A=Ascendent, D=Descendent
    #
    #topsites_sort_order CONNECT D
    
    # TAG: index_sort_order A/D
    #      Sort for index.html, where A=Ascendent, D=Descendent
    #
    #index_sort_order D
    
    # TAG: exclude_codes file
    #      Ignore records with these codes. Eg.: NONE/400
    #      Write one code per line. Lines starting with a # are ignored.
    #      Only codes matching exactly one of the line is rejected. The
    #      comparison is not case sensitive.
    #
    exclude_codes /usr/local/etc/sarg/exclude_codes
    
    # TAG: replace_index string
    #      Replace "index.html" in the main index file with this string
    #      If null "index.html" is used
    #
    #replace_index 
    
    # TAG: max_elapsed milliseconds
    #      If elapsed time is recorded in log is greater than max_elapsed use 0 for elapsed time.
    #      Use 0 for no checking
    #
    #max_elapsed 28800000
    # 8 Hours
    max_elapsed 0
    
    # TAG: report_type type
    #      What kind of reports to generate.
    #      topusers            - users, sites, times, bytes, connects, links to accessed sites, etc
    #      topsites		   - site, connect and bytes report
    #      sites_users	   - users and sites report
    #      users_sites	   - accessed sites by the user report
    #      date_time	   - bytes used per day and hour report
    #      denied		   - denied sites with full URL report
    #      auth_failures       - autentication failures report
    #      site_user_time_date - sites, dates, times and bytes report
    #      downloads           - downloads per user report
    #
    #      Eg.: report_type topsites denied
    #
    #report_type topusers topsites sites_users users_sites date_time denied auth_failures site_user_time_date downloads
    report_type topusers topsites sites_users users_sites date_time denied auth_failures site_user_time_date downloads
    
    # TAG: usertab filename
    #      You can change the "userid" or the "ip address" to be a real user name on the reports.
    #      If resolve_ip is active, the ip address is resolved before being looked up into this
    #      file. That is, if you want to map the ip address, be sure to set resolv_ip to no or
    #      the resolved name will be looked into the file instead of the ip address. Note that
    #      it can be used to resolve any ip address known to the dns and then map the unresolved
    #      ip addresses to a name found in the usertab file.
    #      Table syntax:
    # 		userid name   or   ip address name
    #      Eg:
    #		SirIsaac Isaac Newton
    #		vinci Leonardo da Vinci
    #		192.168.10.1 Karol Wojtyla
    #
    #      Each line must be terminated with '\ n'
    #      If usertab have value "ldap" (case ignoring), user names
    #      will be taken from LDAP server. This method as approaches for reception
    #      of usernames from Active Didectory
    #
    #usertab none
    usertab none
    
    # TAG: LDAPHost hostname
    #	FQDN or IP address of host with LDAP service or AD DC
    #	default is '127.0.0.1'
    #LDAPHost 127.0.0.1
    
    # TAG: LDAPPort port
    #       LDAP service port number
    #	default is '389'
    #LDAPPort 389
    
    # TAG: LDAPBindDN CN=username,OU=group,DC=mydomain,DC=com
    #	DN of LDAP user, who is authorized to read user's names from LDAP base
    #	default is empty line
    #LDAPBindDN cn=proxy,dc=mydomain,dc=local
    
    # TAG: LDAPBindPW secret
    #	Password of DN, who is authorized to read user's names from LDAP base
    #	default is empty line
    #LDAPBindPW secret
    
    # TAG: LDAPBaseSearch OU=users,DC=mydomain,DC=com
    #	LDAP search base
    #	default is empty line
    #LDAPBaseSearch ou=users,dc=mydomain,dc=local
    
    # TAG: LDAPFilterSearch (uid=%s)
    #	User search filter by user's logins in LDAP
    #	First founded record will be used
    #	%s - will be changed to userlogins from access.log file
    #       filter string can have up to 5 '%s' tags
    #	default value is '(uid=%s)'
    #LDAPFilterSearch (uid=%s)
    
    # TAG: LDAPTargetAttr attributename
    #	Name of the attribute containing a name of the user
    #	default value is 'cn'
    #LDAPTargetAttr cn
    
    # TAG: long_url yes|no
    #      If yes, the full url is showed in report.
    #      If no, only the site will be showed
    #
    #      YES option generate very big sort files and reports.
    #
    long_url yes
    
    # TAG: date_time_by bytes|elap
    #      Date/Time reports show the downloaded volume or the elapsed time or both.
    #
    #date_time_by bytes
    date_time_by bytes elap
    
    # TAG: charset name
    #      ISO 8859 is a full series of 10 standardized multilingual single-byte coded (8bit)
    #      graphic character sets for writing in alphabetic languages
    #      You can use the following charsets:
    #		Latin1 		- West European
    #		Latin2 		- East European
    #		Latin3 		- South European
    #		Latin4 		- North European
    #		Cyrillic
    #		Arabic
    #		Greek
    #		Hebrew
    #		Latin5 		- Turkish
    #		Latin6
    #		Windows-1251
    #		Japan
    #		Koi8-r
    #		UTF-8
    #
    #charset Latin1
    charset UTF-8
    
    # TAG: user_invalid_char "&/"
    #      Records that contain invalid characters in userid will be ignored by Sarg.
    #
    #user_invalid_char "&/"
    
    # TAG: privacy yes|no
    #      privacy_string "***.***.***.***"
    #      privacy_string_color blue
    #      In some countries the sysadm cannot see the visited sites by a restrictive law.
    #      Using privacy yes the visited url will be changes by privacy_string and the link
    #      will be removed from reports.
    #
    privacy no
    #privacy_string "***.***.***.***"
    #privacy_string_color blue
    
    # TAG: include_users "user1:user2:...:usern"
    #      Reports will be generated only for listed users.
    #
    #include_users none
    
    # TAG: exclude_string "string1:string2:...:stringn"
    #      Records from access.log file that contain one of listed strings will be ignored.
    #
    #exclude_string none
    
    # TAG: show_successful_message yes|no
    #      Shows "Successful report generated on dir" at end of process.
    #
    #show_successful_message yes
    
    # TAG: show_read_statistics yes|no
    #      Shows some reading statistics.
    #
    #show_read_statistics yes
    
    # TAG: topuser_fields
    #      Which fields must be in Topuser report.
    #
    #topuser_fields NUM DATE_TIME USERID CONNECT BYTES %BYTES IN-CACHE-OUT USED_TIME MILISEC %TIME TOTAL AVERAGE
    
    # TAG: user_report_fields
    #      Which fields must be in User report.
    #
    #user_report_fields CONNECT BYTES %BYTES IN-CACHE-OUT USED_TIME MILISEC %TIME TOTAL AVERAGE
    
    # TAG: bytes_in_sites_users_report yes|no
    #      Bytes field must be in Site & Users Report ?
    #
    #bytes_in_sites_users_report no
    bytes_in_sites_users_report yes
    
    # TAG: topuser_num n
    #      How many users in topsites report. 0 = no limit
    #
    #topuser_num 0
    topuser_num 0
    
    # TAG: datafile file
    #      Save the report results in a file to populate some database
    #
    #datafile none
    
    # TAG: datafile_delimiter "	"
    #      ascii character to use as a field separator in datafile
    #
    #datafile_delimiter ""
    
    # TAG: datafile_fields all
    #      Which data fields must be in datafile
    #      user;date;time;url;connect;bytes;in_cache;out_cache;elapsed
    #
    #datafile_fields user;date;time;url;connect;bytes;in_cache;out_cache;elapsed
    
    # TAG: datafile_url ip|name
    #      Saves the URL as ip or name in datafile
    #
    #datafile_url ip
    
    # TAG: weekdays
    #      The weekdays to take into account ( Sunday->0, Saturday->6 )
    # Example:
    #weekdays 1-3,5
    # Default:
    #weekdays 0-6
    
    # TAG: hours
    #      The hours to take into account
    # Example:
    #hours 7-12,14,16,18-20
    # Default:
    #hours 0-23
    
    # TAG: dansguardian_conf file
    #      DansGuardian.conf file path
    #      Generate reports from DansGuardian logs.
    #      Use 'none' to disable it.
    #      dansguardian_conf /usr/dansguardian/dansguardian.conf
    #
    dansguardian_conf /usr/local/etc/dansguardian/dansguardian.conf
    
    # TAG: dansguardian_filter_out_date on|off
    #      This option replaces dansguardian_ignore_date whose name was not appropriate with respect to its action.
    #      Note the change of parameter value compared with the old option.
    #      'off' use the record even if its date is outside of the range found in the input log file.
    #      'on'  use the record only if its date is in the range found in the input log file.
    #
    dansguardian_filter_out_date on
    
    # TAG: squidguard_conf file
    #      path to squidGuard.conf file
    #      Generate reports from SquidGuard logs.
    #      Use 'none' to disable.
    #      You can use sarg -L filename to use an alternate squidGuard log.
    #      squidguard_conf /usr/local/squidGuard/squidGuard.conf
    #
    squidguard_conf none
    
    # TAG: redirector_log file
    #      the location of the web proxy redirector log such as one created by squidGuard or Rejik. The option
    #      may be repeated up to 64 times to read multiple files.
    #      If this option is specified, it takes precedence over squidguard_conf.
    #      The command line option -L override this option.
    #
    #redirector_log /usr/local/squidGuard/var/logs/urls.log
    
    # TAG: redirector_filter_out_date on|off
    #      This option replaces squidguard_ignore_date and redirector_ignore_date whose names were not
    #      appropriate with respect to their action.
    #      Note the change of parameter value compared with the old options.
    #      'off' use the record even if its date is outside of the range found in the input log file.
    #      'on'  use the record only if its date is in the range found in the input log file.
    #
    #redirector_filter_out_date on
    
    # TAG: redirector_log_format
    #      Format string for web proxy redirector logs.
    #      This option was named squidguard_log_format before sarg 2.3.
    #      REJIK       #year#-#mon#-#day# #hour# #list#:#tmp# #ip# #user# #tmp#/#tmp#/#url#/#end#
    #      SQUIDGUARD  #year#-#mon#-#day# #hour# #tmp#/#list#/#tmp#/#tmp#/#url#/#tmp# #ip#/#tmp# #user# #end#
    #redirector_log_format #year#-#mon#-#day# #hour# #tmp#/#list#/#tmp#/#tmp#/#url#/#tmp# #ip#/#tmp# #user# #end#
    
    # TAG: show_sarg_info yes|no
    #      shows sarg information and site path on each report bottom
    #
    show_sarg_info no
    
    # TAG: show_sarg_logo yes|no
    #      shows sarg logo
    #
    show_sarg_logo no
    
    # TAG: parsed_output_log directory
    #      Saves the processed log in a sarg format after parsing the squid log file.
    #      This is a way to dump all of the data structures out, after parsing from
    #      the logs (presumably this data will be much smaller than the log files themselves),
    #      and pull them back in for later processing and merging with data from previous logs.
    #
    #parsed_output_log none
    
    # TAG: parsed_output_log_compress /bin/gzip|/usr/bin/bzip2|nocompress
    #      Command to run to compress sarg parsed output log. It may contain
    #      options (such as -f to overwrite existing target file). The name of
    #      the file to compresse is provided at the end of this
    #      command line. Don't forget to quote things appropriately.
    #
    #parsed_output_log_compress /bin/gzip
    
    # TAG: displayed_values bytes|abbreviation
    #      how the values will be displayed in reports.
    #      eg. bytes  	-  209.526
    #          abbreviation -  210K
    #
    #displayed_values bytes
    displayed_values abbreviation
    
    # Report limits
    # TAG: authfail_report_limit n
    # TAG: denied_report_limit n
    # TAG: siteusers_report_limit n
    # TAG: squidguard_report_limit n
    # TAG: user_report_limit n
    # TAG: dansguardian_report_limit n
    # TAG: download_report_limit n
    #      report limits (lines).
    #      '0' no limit
    #
    #authfail_report_limit 10
    authfail_report_limit 0
    #denied_report_limit 10
    denied_report_limit 0
    #siteusers_report_limit 0
    #squidguard_report_limit 10
    #dansguardian_report_limit 10
    #user_report_limit 10
    #user_report_limit 50
    siteusers_report_limit 0
    user_report_limit 0
    dansguardian_report_limit 0
    
    # TAG: www_document_root dir
    #     Where is your Web DocumentRoot
    #     Sarg will create sarg-php directory with some PHP modules:
    #     - sarg-squidguard-block.php - add urls from user reports to squidGuard DB
    #
    #www_document_root /var/www/html
    www_document_root /usr/local/www
    
    # TAG: block_it module_url
    #     This tag allow you to pass urls from user reports to a cgi or php module,
    #     to be blocked by some Squid acl
    #
    #     Eg.: block_it /sarg-php/sarg-block-it.php
    #     sarg-block-it is a php that will append a url to a flat file.
    #     You must change /var/www/html/sarg-php/sarg-block-it to point to your file
    #     in  variable, and chown to a httpd owner.
    #
    #     sarg will pass http://module_url?url=url
    #
    #block_it none
    
    # TAG: external_css_file path
    #     Provide the path to an external css file to link into the HTML reports instead of
    #     the inline css written by sarg when this option is not set.
    #
    #     In versions prior to 2.3, this used to be an absolute file name to
    #     a file to include verbatim in each HTML page but, as it takes a lot of
    #     space, version 2.3 switched to a link to an external css file.
    #     Therefore, this option must contain the HTTP server path on which a client
    #     browser may find the css file.
    #
    #     Sarg use theses style classes:
    #	.logo		logo class
    #	.info		sarg information class, align=center
    #	.title_c	title class, align=center
    #	.header_c	header class, align:center
    #	.header_l	header class, align:left
    #	.header_r	header class, align:right
    #	.text		text class, align:right
    #	.data		table text class, align:right
    #	.data2		table text class, align:left
    #	.data3		table text class, align:center
    #	.link  		link class
    #
    #     Sarg can be instructed to output the internal css it inline
    #     into the reports with this command:
    #
    #        sarg --css
    #
    #     You can redirect the output to a file of your choice and edit
    #     it to your liking.
    #
    #external_css_file none
    # TAG: user_authentication yes|no
    #     Allow user authentication in User Reports using .htaccess
    #     Parameters:
    #	AuthUserTemplateFile - The template to use to create the
    #     .htaccess file. In the template, %u is replaced by the
    #     user's ID for which the report is generated. The path of the
    #     template is relative to the directory containing sarg
    #     configuration file.
    #
    # user_authentication no
    # AuthUserTemplateFile sarg_htaccess
    
    # TAG: download_suffix "suffix,suffix,...,suffix"
    #    file suffix to be considered as "download" in Download report.
    #    Use 'none' to disable.
    #
    #download_suffix "zip,arj,bzip,gz,ace,doc,iso,adt,bin,cab,com,dot,drv$,lha,lzh,mdb,mso,ppt,rtf,src,shs,sys,exe,dll,mp3,avi,mpg,mpeg"
    
    # TAG: ulimit n
    #    The maximum number of open file descriptors to avoid "Too many open files" error message.
    #    You need to run sarg as root to use ulimit tag.
    #    If you run sarg with a low privilege user, set to 'none' to disable ulimit
    #
    #ulimit 20000
    
    # TAG: ntlm_user_format username|domainname+username
    #      NTLM users format.
    #
    #ntlm_user_format domainname+username
    ntlm_user_format domainname+username
    
    # TAG: realtime_refresh_time num sec
    #      How many time to auto refresh the realtime report
    #      0 = disable
    #
    realtime_refresh_time 0
    
    # TAG: realtime_access_log_lines num
    #      How many last lines to get from access.log file
    #
    # realtime_access_log_lines 1000
    
    # TAG: realtime_types: GET,PUT,CONNECT,ICP_QUERY,POST
    #      Which records must be in realtime report.
    #
    realtime_types GET,PUT,CONNECT
    
    # TAG: realtime_unauthenticated_records: ignore|show
    #      What to do with unauthenticated records in realtime report.
    #
    # 
    realtime_unauthenticated_records show
    
    # TAG: byte_cost value no_cost_limit
    #      Cost per byte.
    #      Eg. byte_cost 0.01 100000000
    #           per byte cost      = 0.01
    #           bytes with no cost = 100 Mb
    #      0 = disable
    #
    # byte_cost 0.01 50000000
    
    # TAG: squid24 on|off
    #      Compatilibity with squid version <= 2.4 when using emulate_http_log on
    #
    # squid24 off
    
    # TAG: sorttable path
    #      The path to a javascript script to dynamically sort the tables.
    #      The path is the link a browser must follow to find the script. For instance,
    #      it may be http://www.myproxy.org/sorttable.js or just /sorttable.js if the script
    #      is at the root of your web site.
    #
    #      If the path starts with "../" then it is assumed to be a relative
    #      path and sarg adds as many "../" as necessary to locate the js script from
    #      the output directory. Therefore, ../../sorttable.js links to the javascript
    #      one level above output_dir.
    #
    #      If this entry is set, each sortable table will have the "sortable" class set.
    #      You may have a look at http://www.kryogenix.org/code/browser/sorttable/
    #      for the implementation on which sarg is based.
    #
    sorttable /sarg_sorttable.js
    
    # TAG: hostalias
    #      The name of a text file containing the host names one per line and the
    #      optional alias to use in the report instead of that host name.
    #      Host names may contain up to one wildcard denoted by a *. The wildcard
    #      must not end the host name.
    #      The host name may be followed by an optional alias but if no alias is
    #      provided, the host name, including the wildcard, replaces any matching
    #      host name found in the log.
    #      Host names replaced by identical aliases are grouped together in the
    #      reports.
    #      IP addresses are supported and accept the CIDR notation both for IPv4 and
    #      IPv6 addresses.
    #
    #      Example:
    #      *.gstatic.com
    #      mt*.google.com
    #      *.myphone.microsoft.com
    #      *.myphone.microsoft.com:443 *.myphone.microsoft.com:secure
    #      *.freeav.net antivirus:freeav
    #      *.mail.live.com
    #      65.52.00.00/14 *.mail.live.com
    hostalias /usr/local/etc/sarg/hostalias
    

    Log file:

    Jan 21 21:49:47	php: /index.php: Successful webConfigurator login for user 'admin' from ###.###.###.###
    Jan 21 21:00:03	php: : The command 'export LC_ALL=C && /usr/local/bin/sarg -d `date +%d/%m/%Y`' returned exit code '1', the output was 'SARG: Records in file: 6135, reading: 0.00%^MSARG: Records in file: 5000, reading: 81.50%^MSARG: Period covered by log files: 21/01/2014-21/01/2014 sort: open failed: /tmp/sarg/denied.int_unsort: No such file or directory SARG: sort command return status 2 SARG: sort command: sort -T "/tmp/sarg" -t "	" -k 3,3 -k 5,5 -o "/tmp/sarg/denied.int_log" "/tmp/sarg/denied.int_unsort" SARG: Records in file: 6135, reading: 100.00%'
    Jan 21 21:00:03	php: : The command 'export LC_ALL=C && /usr/local/bin/sarg -d `date +%d/%m/%Y`' returned exit code '1', the output was 'SARG: Records in file: 6135, reading: 0.00%^MSARG: Records in file: 5000, reading: 81.50%^MSARG: Period covered by log files: 21/01/2014-21/01/2014 sort: open failed: /tmp/sarg/10_20_10_1.user_unsort: No such file or directory SARG: sort command return status 2 SARG: sort command: sort -T "/tmp/sarg" -t "	" -k 4,4 -k 1,1 -k 2,2 -o "/tmp/sarg/10_20_10_1.user_log" "/tmp/sarg/10_20_10_1.user_unsort" SARG: Records in file: 6135, reading: 100.00%'
    Jan 21 21:00:03	php: : The command 'export LC_ALL=C && /usr/local/bin/sarg -d `date +%d/%m/%Y`' returned exit code '1', the output was 'SARG: Records in file: 6135, reading: 0.00%^MSARG: Records in file: 5000, reading: 81.50%^MSARG: Period covered by log files: 21/01/2014-21/01/2014 sort: open failed: /tmp/sarg/denied.int_unsort: No such file or directory SARG: sort command return status 2 SARG: sort command: sort -T "/tmp/sarg" -t "	" -k 3,3 -k 5,5 -o "/tmp/sarg/denied.int_log" "/tmp/sarg/denied.int_unsort" SARG: Records in file: 6135, reading: 100.00%'
    Jan 21 21:00:03	php: : The command 'export LC_ALL=C && /usr/local/bin/sarg -d `date +%d/%m/%Y`' returned exit code '1', the output was 'SARG: Records in file: 6135, reading: 0.00%^MSARG: Records in file: 5000, reading: 81.50%^MSARG: Period covered by log files: 21/01/2014-21/01/2014 sort: open failed: /tmp/sarg/denied.int_unsort: No such file or directory SARG: sort command return status 2 SARG: sort command: sort -T "/tmp/sarg" -t "	" -k 3,3 -k 5,5 -o "/tmp/sarg/denied.int_log" "/tmp/sarg/denied.int_unsort" SARG: Records in file: 6135, reading: 100.00%'
    Jan 21 21:00:02	php: : The command 'export LC_ALL=C && /usr/local/bin/sarg -d `date +%d/%m/%Y`' returned exit code '1', the output was 'SARG: Records in file: 6135, reading: 0.00%^MSARG: Records in file: 5000, reading: 81.50%^MSARG: Period covered by log files: 21/01/2014-21/01/2014 sort: open failed: /tmp/sarg/denied.int_unsort: No such file or directory SARG: sort command return status 2 SARG: sort command: sort -T "/tmp/sarg" -t "	" -k 3,3 -k 5,5 -o "/tmp/sarg/denied.int_log" "/tmp/sarg/denied.int_unsort" SARG: Records in file: 6135, reading: 100.00%'
    Jan 21 21:00:02	php: : Sarg: force refresh now with -d `date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 21 21:00:02	php: : Sarg: force refresh now with -d `date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 21 21:00:02	php: : Sarg: force refresh now with -d `date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 21 21:00:02	php: : The command 'export LC_ALL=C && /usr/local/bin/sarg -d `date +%d/%m/%Y`' returned exit code '1', the output was 'SARG: cannot stat "/tmp/sarg/authfail.int_unsort"'
    Jan 21 21:00:02	php: : Sarg: force refresh now with -d `date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 21 21:00:02	php: : Sarg: force refresh now with -d `date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 21 21:00:02	php: : Sarg: force refresh now with -d `date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 21 21:00:02	php: : Sarg: force refresh now with -d `date +%d/%m/%Y` args, compress(on) and none action after sarg finish.
    Jan 21 20:55:18	root: rc.update_bogons.sh is sleeping for 10503
    Jan 21 20:55:18	root: rc.update_bogons.sh is starting up.
    Jan 21 20:55:18	root: rc.update_bogons.sh is ending the update cycle
    


  • What about:

    Sarg log
    /tmp/php_errors.txt
    crontab



  • sarg.log - I can't seem to find this, what path should I be looking in?

    /tmp/php_errors.txt = empty file

    crontab:

    
    #
    # pfSense specific crontab entries
    # Created: January 21, 2014, 9:51 am
    #
    
    0	*	*	*	*	root	/usr/bin/nice -n20 newsyslog
    1,31	0-5	*	*	*	root	/usr/bin/nice -n20 adjkerntz -a
    1	3	1	*	*	root	/usr/bin/nice -n20 /etc/rc.update_bogons.sh
    */60	*	*	*	*	root	/usr/bin/nice -n20 /usr/local/sbin/expiretable -v -t 3600 sshlockout
    1	1	*	*	*	root	/usr/bin/nice -n20 /etc/rc.dyndns.update
    */60	*	*	*	*	root	/usr/bin/nice -n20 /usr/local/sbin/expiretable -v -t 3600 virusprot
    30	12	*	*	*	root	/usr/bin/nice -n20 /etc/rc.update_urltables
    0	0	*	*	*	root	/bin/rm /var/squid/cache/swap.state; /usr/local/sbin/squid -k rotate -f /usr/local/etc/squid/squid.conf
    */15	*	*	*	*	root	/usr/local/pkg/swapstate_check.php
    0	5	*	*	*	root	/usr/local/bin/mail_reports_generate.php 0 &
    0	*/2	*	*	*	root	/usr/bin/perl /usr/local/www/lightsquid/lightparser.pl today
    15	0	*	*	*	root	/usr/bin/perl /usr/local/www/lightsquid/lightparser.pl yesterday
    0	*/1	*	*	*	root	/usr/local/bin/php /usr/local/www/sarg.php 0
    #
    # If possible do not add items to this file manually.
    # If you do so, this file must be terminated with a blank line (e.g. new line)
    #
    


  • Does the /tmp/sarg/ directory exist? What files are in there?

    Can you post contents of sarg.php that is called in your crontab?



  • @thestealth:

    Sure no problem. Just a note, when I manually do a force refresh everything works just fine.

    How are you forcing manually



  • @bryan.paradis:

    Does the /tmp/sarg/ directory exist? What files are in there?

    Can you post contents of sarg.php that is called in your crontab?

    There is no /tmp/sarg/

    I do a force refresh by doing Status -> Sarg Reports -> Schedule -> Edit this item -> Click the "Force update now" button.

    Instead of using the built in scheduling tool, could I just create a custom cron job to pass the additional args to Sarg? If so what is the form of the command?

    sarg.php:

    /* $Id$ */
    /* ========================================================================== */
    /*
        sarg.php
        part of pfSense (http://www.pfSense.com)
        Copyright (C) 2012 Marcello Coutinho
    
        All rights reserved.                                                      
    */
    /* ========================================================================== */
    /*
        Redistribution and use in source and binary forms, with or without
        modification, are permitted provided that the following conditions are met:
    
         1\. Redistributions of source code must retain the above copyright notice,
            this list of conditions and the following disclaimer.
    
         2\. Redistributions in binary form must reproduce the above copyright
            notice, this list of conditions and the following disclaimer in the
            documentation and/or other materials provided with the distribution.
    
        THIS SOFTWARE IS PROVIDED ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES,
        INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY
        AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
        AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY,
        OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
        SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
        INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
        CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
        ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
        POSSIBILITY OF SUCH DAMAGE.
                                                                                  */
    /* ========================================================================== */
    
    require_once("/etc/inc/util.inc");
    require_once("/etc/inc/functions.inc");
    require_once("/etc/inc/pkg-utils.inc");
    require_once("/etc/inc/globals.inc");
    require_once("/usr/local/pkg/sarg.inc");
    
    $uname=posix_uname();
    if ($uname['machine']=='amd64')
            ini_set('memory_limit', '250M');
    
    if (preg_match ("/(\d+)/",$argv[1],$matches))
    	run_sarg($matches[1]);
    
    ?>
    


  • Crontab

    
    0	*/1	*	*	*	root	/usr/local/bin/php /usr/local/www/sarg.php 0
    
    

    Cron starts the job

    
    //sarg.php
    if (preg_match ("/(\d+)/",$argv[1],$matches))
            run_sarg($matches[1]);
    
    

    sarg.php  takes out the "0" and passes it to run_sarg("0")

    
    //sarg.inc
    function run_sarg($id=-1) {
            global $config, $g,$sarg_proxy;
            #mount filesystem writeable
            conf_mount_rw();
            $cmd =  SARG_DIR . "/bin/sarg";
            if ($id >= 0 && is_array($config['installedpackages']['sargschedule']['config'])){
                    $args=$config['installedpackages']['sargschedule']['config'][$id]['args'];
                    $action=$config['installedpackages']['sargschedule']['config'][$id]['action'];
                    $gzip=$config['installedpackages']['sargschedule']['config'][$id]['gzip'];
                    $find=$config['installedpackages']['sargschedule']['config'][$id]['find'];
                    $gziplevel=$config['installedpackages']['sargschedule']['config'][$id]['gziplevel'];
                    $daylimit=$config['installedpackages']['sargschedule']['config'][$id]['daylimit'];
                    }
            else{
                    $args=$_POST['args'];
                    $action=$_POST['action'];
                    $gzip=$_POST['gzip'];
                    $find=$_POST['find'];
                    $gziplevel=$_POST['gziplevel'];
                    $daylimit="";
                    }
    
    

    The  if ($id >= 0 && is_array($config['installedpackages']['sargschedule']['config']) should get triggered as $id should = 0

    Force Update Now

    
    //sarg.inc
    if ($_POST['Submit'] == 'Force update now')
                    run_sarg();
    
    

    Pressing the button launches off sarg but without the php variable being passed to the function

    
    //sarg.inc
    function run_sarg($id=-1) {
            global $config, $g,$sarg_proxy;
            #mount filesystem writeable
            conf_mount_rw();
            $cmd =  SARG_DIR . "/bin/sarg";
            if ($id >= 0 && is_array($config['installedpackages']['sargschedule']['config'])){
                    $args=$config['installedpackages']['sargschedule']['config'][$id]['args'];
                    $action=$config['installedpackages']['sargschedule']['config'][$id]['action'];
                    $gzip=$config['installedpackages']['sargschedule']['config'][$id]['gzip'];
                    $find=$config['installedpackages']['sargschedule']['config'][$id]['find'];
                    $gziplevel=$config['installedpackages']['sargschedule']['config'][$id]['gziplevel'];
                    $daylimit=$config['installedpackages']['sargschedule']['config'][$id]['daylimit'];
                    }
            else{
                    $args=$_POST['args'];
                    $action=$_POST['action'];
                    $gzip=$_POST['gzip'];
                    $find=$_POST['find'];
                    $gziplevel=$_POST['gziplevel'];
                    $daylimit="";
                    }
    
    

    Default value id=-1 is used and the ELSE is triggered and I think the schedule args are taken correctly from the POST variables. I wonder if there is a problem

    Going forward

    1. Could you please copy and paste the installed packages section of your /cf/conf/config.xml ? Like below except yours will have sarg etc

    	 <installedpackages><tab><menu>
    		 <service><package><name>vnstat2</name>
    			<website>http://humdi.net/vnstat/</website>
    
    			<pkginfolink>http://forum.pfsense.org/index.php/topic,14179.0.html</pkginfolink>
    			<category>Network Management</category>
    			<depends_on_package_base_url>http://192.168.55.4/packages/8/All/</depends_on_package_base_url>
    			<depends_on_package>vnstat-1.11.tbz</depends_on_package>
    			<depends_on_package_pbi>vnstat-1.11_1-i386.pbi</depends_on_package_pbi>
    			<build_port_path>/usr/ports/net/vnstat</build_port_path>
    			<version>1.10_2</version>
    			<status>Stable</status>
    			<required_version>2.0</required_version>
    			<maintainer>crazypark2@yahoo.dk</maintainer>
    			<config_file>http://www.pfsense.com/packages/config/vnstat2/vnstat2.xml</config_file>
    			<configurationfile>vnstat2.xml</configurationfile>
    			 <after_install_info></after_install_info></package> 
    
    2\. Please check to make sure you do not having multiple cron processes running
    
    

    ps aux | grep cron



  • 1. Too big to fit here, see link: www.spiralex.net/images/packages.xml

    ps aux | grep cron
    root     241  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root     819  0.0  0.2  3348   760  ??  INs  30Nov13   0:06.38 /usr/sbin/cron -s
    root   13181  0.0  0.3  3348  1308  ??  INs  Sat11PM   0:00.48 /usr/sbin/cron -s
    root   16048  0.0  0.3  3348  1308  ??  INs  Sat11PM   0:00.46 /usr/sbin/cron -s
    root   17501  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 240 /var/run/ping_hosts.pid /usr/local/bin/ping_hosts.sh
    root   17640  0.0  0.0  3256   236  ??  I    18Nov13   0:05.04 minicron: helper /usr/local/bin/ping_hosts.sh  (minicron)
    root   17958  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 3600 /var/run/expire_accounts.pid /etc/rc.expireaccounts
    root   18383  0.0  0.0  3256   236  ??  I    18Nov13   0:00.37 minicron: helper /etc/rc.expireaccounts  (minicron)
    root   18565  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 86400 /var/run/update_alias_url_data.pid /etc/rc.update_alias_url_data
    root   18866  0.0  0.0  3256   236  ??  I    18Nov13   0:00.02 minicron: helper /etc/rc.update_alias_url_data  (minicron)
    root   21905  0.0  0.2  3348   760  ??  INs  18Nov13   0:07.80 /usr/sbin/cron -s
    root   28740  0.0  0.2  3348   760  ??  INs   8Dec13   0:05.44 /usr/sbin/cron -s
    root   29429  0.0  0.2  3348   760  ??  INs  11Jan14   0:01.31 /usr/sbin/cron -s
    root   35786  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.01 cron: running job (cron)
    root   36064  0.0  0.3  3348  1304  ??  IN    1Dec13   0:00.00 cron: running job (cron)
    root   37549  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.00 cron: running job (cron)
    root   42116  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.01 cron: running job (cron)
    root   44997  0.0  0.2  3348   760  ??  INs  30Nov13   0:06.40 /usr/sbin/cron -s
    root   54704  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   57223  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   59479  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   61592  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   24039  0.0  0.2  3460  1188   0  S+    6:01PM   0:00.00 grep cron
    
    


  • @thestealth:

    1. Too big to fit here, see link: www.spiralex.net/images/packages.xml

    Looks fine.

    ps aux | grep cron
    root     241  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root     819  0.0  0.2  3348   760  ??  INs  30Nov13   0:06.38 /usr/sbin/cron -s
    root   13181  0.0  0.3  3348  1308  ??  INs  Sat11PM   0:00.48 /usr/sbin/cron -s
    root   16048  0.0  0.3  3348  1308  ??  INs  Sat11PM   0:00.46 /usr/sbin/cron -s
    root   17501  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 240 /var/run/ping_hosts.pid /usr/local/bin/ping_hosts.sh
    root   17640  0.0  0.0  3256   236  ??  I    18Nov13   0:05.04 minicron: helper /usr/local/bin/ping_hosts.sh  (minicron)
    root   17958  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 3600 /var/run/expire_accounts.pid /etc/rc.expireaccounts
    root   18383  0.0  0.0  3256   236  ??  I    18Nov13   0:00.37 minicron: helper /etc/rc.expireaccounts  (minicron)
    root   18565  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 86400 /var/run/update_alias_url_data.pid /etc/rc.update_alias_url_data
    root   18866  0.0  0.0  3256   236  ??  I    18Nov13   0:00.02 minicron: helper /etc/rc.update_alias_url_data  (minicron)
    root   21905  0.0  0.2  3348   760  ??  INs  18Nov13   0:07.80 /usr/sbin/cron -s
    root   28740  0.0  0.2  3348   760  ??  INs   8Dec13   0:05.44 /usr/sbin/cron -s
    root   29429  0.0  0.2  3348   760  ??  INs  11Jan14   0:01.31 /usr/sbin/cron -s
    root   35786  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.01 cron: running job (cron)
    root   36064  0.0  0.3  3348  1304  ??  IN    1Dec13   0:00.00 cron: running job (cron)
    root   37549  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.00 cron: running job (cron)
    root   42116  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.01 cron: running job (cron)
    root   44997  0.0  0.2  3348   760  ??  INs  30Nov13   0:06.40 /usr/sbin/cron -s
    root   54704  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   57223  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   59479  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   61592  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   24039  0.0  0.2  3460  1188   0  S+    6:01PM   0:00.00 grep cron
    
    

    Wish we had checked for multiple cron processes earlier. Learnt some new things though  ;)

    Your problem is with a cron package from early or before october. Please upgrade your cron package and if you have to stop the cron service or kill them off and start the cron service.

    https://forum.pfsense.org/index.php?topic=67445.0



  • @bryan.paradis:

    Wish we had checked for multiple cron processes earlier. Learnt some new things though  ;)

    Your problem is with a cron package from early or before october. Please upgrade your cron package and if you have to stop the cron service or kill them off and start the cron service.

    https://forum.pfsense.org/index.php?topic=67445.0

    This seems to have worked:

    ps aux | grep cron
    root     241  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   17501  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 240 /var/run/ping_hosts.pid /usr/local/bin/ping_hosts.sh
    root   17640  0.0  0.0  3256   236  ??  I    18Nov13   0:05.05 minicron: helper /usr/local/bin/ping_hosts.sh  (minicron)
    root   17958  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 3600 /var/run/expire_accounts.pid /etc/rc.expireaccounts
    root   18383  0.0  0.0  3256   236  ??  I    18Nov13   0:00.37 minicron: helper /etc/rc.expireaccounts  (minicron)
    root   18565  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 86400 /var/run/update_alias_url_data.pid /etc/rc.update_alias_url_data
    root   18866  0.0  0.0  3256   236  ??  I    18Nov13   0:00.02 minicron: helper /etc/rc.update_alias_url_data  (minicron)
    root   28032  0.0  0.3  3348  1332  ??  Ss    6:39PM   0:00.01 /usr/sbin/cron -s
    root   35786  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.01 cron: running job (cron)
    root   36064  0.0  0.3  3348  1304  ??  IN    1Dec13   0:00.00 cron: running job (cron)
    root   37549  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.00 cron: running job (cron)
    root   42116  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.01 cron: running job (cron)
    root   54704  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   57223  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   59479  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   61592  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   29000  0.0  0.2  3460  1188   0  S+    7:25PM   0:00.00 grep cron
    
    

    What do you think?



  • @thestealth:

    @bryan.paradis:

    Wish we had checked for multiple cron processes earlier. Learnt some new things though  ;)

    Your problem is with a cron package from early or before october. Please upgrade your cron package and if you have to stop the cron service or kill them off and start the cron service.

    https://forum.pfsense.org/index.php?topic=67445.0

    This seems to have worked:

    ps aux | grep cron
    root     241  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   17501  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 240 /var/run/ping_hosts.pid /usr/local/bin/ping_hosts.sh
    root   17640  0.0  0.0  3256   236  ??  I    18Nov13   0:05.05 minicron: helper /usr/local/bin/ping_hosts.sh  (minicron)
    root   17958  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 3600 /var/run/expire_accounts.pid /etc/rc.expireaccounts
    root   18383  0.0  0.0  3256   236  ??  I    18Nov13   0:00.37 minicron: helper /etc/rc.expireaccounts  (minicron)
    root   18565  0.0  0.0  3256     0  ??  IWs  -         0:00.00 /usr/local/bin/minicron 86400 /var/run/update_alias_url_data.pid /etc/rc.update_alias_url_data
    root   18866  0.0  0.0  3256   236  ??  I    18Nov13   0:00.02 minicron: helper /etc/rc.update_alias_url_data  (minicron)
    root   28032  0.0  0.3  3348  1332  ??  Ss    6:39PM   0:00.01 /usr/sbin/cron -s
    root   35786  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.01 cron: running job (cron)
    root   36064  0.0  0.3  3348  1304  ??  IN    1Dec13   0:00.00 cron: running job (cron)
    root   37549  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.00 cron: running job (cron)
    root   42116  0.0  0.3  3348  1296  ??  IN    1Dec13   0:00.01 cron: running job (cron)
    root   54704  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   57223  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   59479  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   61592  0.0  0.3  3348  1308  ??  IN    1Jan14   0:00.00 cron: running job (cron)
    root   29000  0.0  0.2  3460  1188   0  S+    7:25PM   0:00.00 grep cron
    
    

    What do you think?

    Only 1 now not 6 :) Looks better!



  • I assume I was looking for this line:

    root   28032  0.0  0.3  3348  1332  ??  Ss    6:39PM   0:00.01 /usr/sbin/cron -s
    

    No more weirdness in my log either.

    Thank a lot it is much appreciated.



  • @thestealth:

    I assume I was looking for this line:

    root   28032  0.0  0.3  3348  1332  ??  Ss    6:39PM   0:00.01 /usr/sbin/cron -s
    

    No more weirdness in my log either.

    Thank a lot it is much appreciated.

    Yes as you can see their were 6 running before and you are welcome!


Log in to reply