Without using sed
or awk
, only cut
, how do I get the last field when the number of fields are unknown or change with every line?
cut
command :)? why not any other Linux commands?
sed
or awk
: perl -pe 's/^.+\s+([^\s]+)$/$1/'
.
cut
supports something it doesn't. But I thought it was useful, in that it forces the reader to consider code that's easier to follow. I wanted a quick, simple way to use cut
without needing to use multiple syntaxes for awk
, grep
, sed
, etc. The rev
thing did the trick; very elegant, and something I've never considered (even if clunky for other situations). I also liked reading the other approaches from the other answers.
find | cut -d. -f<last>
is the natural inclination
You could try something like this:
echo 'maps.google.com' | rev | cut -d'.' -f 1 | rev
Explanation
rev reverses "maps.google.com" to be moc.elgoog.spam
cut uses dot (ie '.') as the delimiter, and chooses the first field, which is moc
lastly, we reverse it again to get com
Use a parameter expansion. This is much more efficient than any kind of external command, cut
(or grep
) included.
data=foo,bar,baz,qux
last=${data##*,}
See BashFAQ #100 for an introduction to native string manipulation in bash.
while IFS= read -ra array_var; do :;done <(cmd)
to process a few lines. But for a big file, rev|cut|rev is probably faster! (And of course awk will be faster than that.)
bash
script (assuming you're already using a bash
script). No need to call anything external.
rev
is specific to whatever OS you're using that provides it -- it's not standardized across all UNIX systems. See the chapter listing for the POSIX section on commands and utilities -- it's not there. And ${var##prefix_pattern}
is not in fact bash-specific; it's in the POSIX sh standard, see the end of section 2.6.2 (linked), so unlike rev
, it's always available on any compliant shell.
It is not possible using just cut
. Here is a way using grep
:
grep -o '[^,]*$'
Replace the comma for other delimiters.
Explanation:
-o (--only-matching) only outputs the part of the input that matches the pattern (the default is to print the entire line if it contains a match).
[^,] is a character class that matches any character other than a comma.
* matches the preceding pattern zero or more time, so [^,]* matches zero or more non‑comma characters.
$ matches the end of the string.
Putting this together, the pattern matches zero or more non-comma characters at the end of the string.
When there are multiple possible matches, grep prefers the one that starts earliest. So the entire last field will be matched.
Full example:
If we have a file called data.csv containing
one,two,three
foo,bar
then grep -o '[^,]*$' < data.csv
will output
three
bar
grep -o '^.*,'
rev
add an issue multibyte unicode characters in my case.
sed 's/^.*,//'
which replaces all characters up to and including the last comma with an empty string.
Without awk ?... But it's so simple with awk:
echo 'maps.google.com' | awk -F. '{print $NF}'
AWK is a way more powerful tool to have in your pocket. -F if for field separator NF is the number of fields (also stands for the index of the last)
cut
to achieve the OP's final output is like using a spoon to "cut" steak (pun intended :) ) . awk
is the steak knife.
echo
that may slow down script for long files using awk -F. '{print $NF}' <<< 'maps.google.com'
.
There are multiple ways. You may use this too.
echo "Your string here"| tr ' ' '\n' | tail -n1
> here
Obviously, the blank space input for tr command should be replaced with the delimiter you need.
This is the only solution possible for using nothing but cut:
echo "s.t.r.i.n.g." | cut -d'.' -f2- [repeat_following_part_forever_or_until_out_of_memory:] | cut -d'.' -f2-
Using this solution, the number of fields can indeed be unknown and vary from time to time. However as line length must not exceed LINE_MAX characters or fields, including the new-line character, then an arbitrary number of fields can never be part as a real condition of this solution.
Yes, a very silly solution but the only one that meets the criterias I think.
cut -f2-
in a loop until the output no longer changes.
cut -f2-
until it no longer changes. Otherwise you'd have to buffer the entire file.
If your input string doesn't contain forward slashes then you can use basename
and a subshell:
$ basename "$(echo 'maps.google.com' | tr '.' '/')"
This doesn't use sed
or awk
but it also doesn't use cut
either, so I'm not quite sure if it qualifies as an answer to the question as its worded.
This doesn't work well if processing input strings that can contain forward slashes. A workaround for that situation would be to replace forward slash with some other character that you know isn't part of a valid input string. For example, the pipe (|
) character is also not allowed in filenames, so this would work:
$ basename "$(echo 'maps.google.com/some/url/things' | tr '/' '|' | tr '.' '/')" | tr '|' '/'
touch \|
.
|
being not allowed in file names. But almost every tr
out there supports \0
or some other way of expressing the nul byte, and that definitely isn't allowed in file names, so you can use that as a place holder. Also tr ab bc
just swaps all a
and b
without problems, so you can just avoid having to find a disallowed character entirely. Just pipe through tr './' './'
once to swap before the basename
and then again to swap back after.
tr '/.' './'
once to swap before the basename and then again after".
the following implements A friend's suggestion
#!/bin/bash
rcut(){
nu="$( echo $1 | cut -d"$DELIM" -f 2- )"
if [ "$nu" != "$1" ]
then
rcut "$nu"
else
echo "$nu"
fi
}
$ export DELIM=.
$ rcut a.b.c.d
d
echo
in order for this to work reliably and robustly. See stackoverflow.com/questions/10067266/…
An alternative using perl would be:
perl -pe 's/(.*) (.*)$/$2/' file
where you may change \t
for whichever the delimiter of file
is
If you have a file named filelist.txt that is a list paths such as the following: c:/dir1/dir2/file1.h c:/dir1/dir2/dir3/file2.h
then you can do this: rev filelist.txt | cut -d"/" -f1 | rev
Adding an approach to this old question just for the fun of it:
$ cat input.file # file containing input that needs to be processed
a;b;c;d;e
1;2;3;4;5
no delimiter here
124;adsf;15454
foo;bar;is;null;info
$ cat tmp.sh # showing off the script to do the job
#!/bin/bash
delim=';'
while read -r line; do
while [[ "$line" =~ "$delim" ]]; do
line=$(cut -d"$delim" -f 2- <<<"$line")
done
echo "$line"
done < input.file
$ ./tmp.sh # output of above script/processed input file
e
5
no delimiter here
15454
info
Besides bash, only cut is used. Well, and echo, I guess.
while read -r line; do echo ${line/*;}; done <input.file
yields the same result.
I realized if we just ensure a trailing delimiter exists, it works. So in my case I have comma and whitespace delimiters. I add a space at the end;
$ ans="a, b"
$ ans+=" "; echo ${ans} | tr ',' ' ' | tr -s ' ' | cut -d' ' -f2
b
ans="a, b, c"
produces b
, which does not meet the requirements of "number of fields are unknown or change with every line".
Success story sharing
cut
but it's withoutsed
orawk
.So what OP think?rev
afterwards.rev
great ideal!