I met with a friend the other day, who also happens to be my dental hygienist, and we were talking about work. She said that during the holidays she had developed a back pain but that it really didn't matter because it always happened to her - and that she knew that when she got back to work her "pain would just go away".
Work can be more than just growing professionally and having a good salary. We can find work that also helps us grow physically, emotionally, and even spiritually. Isn't this type of job one worth searching for?đđ
Is it possible for you to imagine a job that is much more that professional growth? Have you ever had a job that has made you feel more healthier, safer, and happier? I would love to hear your thoughts on this in the comments!